Skip to content
GitLab
Explore
Projects
Groups
Topics
Snippets
Projects
Groups
Topics
Snippets
/
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
Menu
MKCLOS
Core Development Platform
corepkgv2
Commits
342e46f8
Commit
342e46f8
authored
5 years ago
by
Vivek Naik
Browse files
Options
Downloads
Patches
Plain Diff
removed commented code
parent
36f3d0b6
2 merge requests
!169
Mep release 060320
,
!167
fdb fixes
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
dalmdl/corefdb/bucket.go
+1
-205
dalmdl/corefdb/bucket.go
dalmdl/corefdb/indexStore.go
+0
-2
dalmdl/corefdb/indexStore.go
with
1 addition
and
207 deletions
dalmdl/corefdb/bucket.go
+
1
−
205
View file @
342e46f8
...
@@ -145,11 +145,6 @@ func (fdb *FDB) NewBucket(bucket, parentBucket *Bucket) *Bucket {
...
@@ -145,11 +145,6 @@ func (fdb *FDB) NewBucket(bucket, parentBucket *Bucket) *Bucket {
return
bucket
return
bucket
}
}
// // SetIsBucketPackType - SetIsBucketPackType
// func (bucket *Bucket) SetIsBucketPackType(isPackType bool) {
// bucket.IsSinglePackType = isPackType
// }
// SetBucketType -
// SetBucketType -
func
(
bucket
*
Bucket
)
SetBucketType
(
bucketType
string
)
{
func
(
bucket
*
Bucket
)
SetBucketType
(
bucketType
string
)
{
bucket
.
BucketType
=
bucketType
bucket
.
BucketType
=
bucketType
...
@@ -318,35 +313,6 @@ func AppendDataInLazyObjectInCache(bucketID string, data gjson.Result, filePath
...
@@ -318,35 +313,6 @@ func AppendDataInLazyObjectInCache(bucketID string, data gjson.Result, filePath
return
nil
return
nil
}
}
// // restoreFilesFromFilePack - restoreFilesFromFilePack
// func restoreFilesFromFilePack(fileName string, fileData []byte) error {
// if len(fileData) > 0 {
// err := filemdl.WriteFile(fileName, fileData, true, false)
// if errormdl.CheckErr(err) != nil {
// loggermdl.LogError(err)
// return errormdl.CheckErr(err)
// }
// }
// return nil
// }
// func readNormalFile(path string) ([]byte, error) {
// ba, err := filemdl.ReadFile(path)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// if isSecurityEnabled {
// _, fileName := filepath.Split(path)
// ba, err = decryptData(ba, fileName)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// }
// return ba, nil
// }
func
readNormalFileUsingFp
(
fp
*
os
.
File
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
([]
byte
,
error
)
{
func
readNormalFileUsingFp
(
fp
*
os
.
File
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
([]
byte
,
error
)
{
ba
,
err
:=
filemdl
.
ReadFileUsingFp
(
fp
)
ba
,
err
:=
filemdl
.
ReadFileUsingFp
(
fp
)
if
err
!=
nil
{
if
err
!=
nil
{
...
@@ -374,37 +340,6 @@ func readNormalFileUsingFp(fp *os.File, rs *gjson.Result, secParams securitymdl.
...
@@ -374,37 +340,6 @@ func readNormalFileUsingFp(fp *os.File, rs *gjson.Result, secParams securitymdl.
return
ba
,
nil
return
ba
,
nil
}
}
// // readNormalFiles - reads file and returns file data
// func (fdb *FDB) readNormalFiles(path []string) (*gjson.Result, error) {
// sb := strings.Builder{}
// sb.WriteString("[")
// for i := 0; i < len(path); i++ {
// updatedPath := filepath.Join(fdb.DBPath, path[i])
// ba := []byte{'{', '}'}
// var err error
// if filemdl.FileAvailabilityCheck(updatedPath) {
// ba, err = readNormalFile(updatedPath)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// // loggermdl.LogError("File data for index = ", i, " - ", string(ba))
// _, err = sb.Write(ba)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// sb.WriteString(",")
// }
// }
// sb.WriteString("]")
// finalResult := strings.Replace(sb.String(), ",]", "]", 1)
// rs := gjson.Parse(finalResult)
// return &rs, nil
// }
// readNormalFilesUsingFp - reads file and returns file data
func
readNormalFilesUsingFp
(
fps
[]
*
os
.
File
,
bucket
*
Bucket
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
(
*
gjson
.
Result
,
error
)
{
func
readNormalFilesUsingFp
(
fps
[]
*
os
.
File
,
bucket
*
Bucket
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
(
*
gjson
.
Result
,
error
)
{
sb
:=
strings
.
Builder
{}
sb
:=
strings
.
Builder
{}
...
@@ -428,7 +363,6 @@ func readNormalFilesUsingFp(fps []*os.File, bucket *Bucket, rs *gjson.Result, se
...
@@ -428,7 +363,6 @@ func readNormalFilesUsingFp(fps []*os.File, bucket *Bucket, rs *gjson.Result, se
loggermdl
.
LogError
(
err
)
loggermdl
.
LogError
(
err
)
return
nil
,
err
return
nil
,
err
}
}
// loggermdl.LogError("File data for index = ", i, " - ", string(ba))
_
,
err
=
sb
.
Write
(
ba
)
_
,
err
=
sb
.
Write
(
ba
)
if
err
!=
nil
{
if
err
!=
nil
{
loggermdl
.
LogError
(
err
)
loggermdl
.
LogError
(
err
)
...
@@ -716,28 +650,6 @@ func deleteDataFromLazyBucket(bucket *Bucket, fp *os.File) error {
...
@@ -716,28 +650,6 @@ func deleteDataFromLazyBucket(bucket *Bucket, fp *os.File) error {
return
nil
return
nil
}
}
// func saveDataInNormalBucketUsingFpNew(f *os.File, bucket *Bucket, rs *gjson.Result) error {
// dataBytes := []byte(rs.String())
// var err error
// if bucket.EnableLazy {
// return saveDatInLazyBucket(bucket, f, rs)
// }
// if isSecurityEnabled {
// _, fileName := filepath.Split(f.Name())
// dataBytes, err = encryptData(dataBytes, fileName)
// if err != nil {
// loggermdl.LogError(err)
// return err
// }
// }
// err = filemdl.WriteFileUsingFp(f, dataBytes, true, false)
// if errormdl.CheckErr(err) != nil {
// loggermdl.LogError(err)
// return errormdl.CheckErr(err)
// }
// return nil
// }
func
saveDataInNormalBucketUsingFp
(
f
*
os
.
File
,
bucket
*
Bucket
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
error
{
func
saveDataInNormalBucketUsingFp
(
f
*
os
.
File
,
bucket
*
Bucket
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
error
{
dataBytes
:=
[]
byte
(
rs
.
String
())
dataBytes
:=
[]
byte
(
rs
.
String
())
var
err
error
var
err
error
...
@@ -774,66 +686,6 @@ func saveDataInNormalBucketUsingFp(f *os.File, bucket *Bucket, rs *gjson.Result,
...
@@ -774,66 +686,6 @@ func saveDataInNormalBucketUsingFp(f *os.File, bucket *Bucket, rs *gjson.Result,
return
nil
return
nil
}
}
// func saveDataInNormalBucket(filePath string, rs *gjson.Result) error {
// dataBytes := []byte(rs.String())
// var err error
// if isSecurityEnabled {
// _, fileName := filepath.Split(filePath)
// dataBytes, err = encryptData(dataBytes, fileName)
// if err != nil {
// loggermdl.LogError(err)
// return err
// }
// }
// err = filemdl.WriteFile(filePath, dataBytes, true, false)
// if errormdl.CheckErr(err) != nil {
// loggermdl.LogError(err)
// return errormdl.CheckErr(err)
// }
// return nil
// }
// func saveDataInPackBucket(fdb *FDB, filePath string, index *Index, rs *gjson.Result) error {
// // check is file locked
// // if yes
// // wait
// // else
// // /* */update filestatus to fileStatusUpdatingData
// // compress and encrypt data
// // calculate filehash
// //save data in file & getoffset and len
// //update footer address
// // update filestatuse to fileStatusUpdatingIndex
// // update index
// // update filestatus to fileStatusReady
// bucketID := index.BucketSequence[len(index.BucketSequence)-1]
// bucket := fdb.buckets[bucketID]
// fileType := rs.Get("fileType").String()
// if len(fileType) == 0 {
// return errormdl.Wrap("please specify fileType")
// }
// isFilePresent := filemdl.FileAvailabilityCheck(filePath)
// infileIndex, ok := bucket.InFileIndexMap[fileType]
// if !ok {
// loggermdl.LogError("infileIndex for specified fileType not found: ", fileType)
// return errormdl.Wrap("infileIndex for specified fileType not found: " + fileType)
// }
// err := filepack.AddFileInPackFile(filePath, infileIndex.IndexFields, rs)
// if err != nil {
// loggermdl.LogError("fail to add fdb index entry in file: ", err)
// return errormdl.Wrap("fail to write data in file")
// }
// // for adding fdb index data
// if !isFilePresent {
// err = addFDBIndexEntryFile(filePath, bucket, index.IndexFields, rs)
// }
// return err
// }
func
saveDataInPackBucketUsingFp
(
fdb
*
FDB
,
f
*
os
.
File
,
index
*
Index
,
infileIndexData
,
rs
*
gjson
.
Result
)
(
*
gjson
.
Result
,
error
)
{
func
saveDataInPackBucketUsingFp
(
fdb
*
FDB
,
f
*
os
.
File
,
index
*
Index
,
infileIndexData
,
rs
*
gjson
.
Result
)
(
*
gjson
.
Result
,
error
)
{
// check is file locked
// check is file locked
// if yes
// if yes
...
@@ -944,11 +796,6 @@ func SaveMediaInFDB(dbName string, indexID string, mediaData []byte, rs *gjson.R
...
@@ -944,11 +796,6 @@ func SaveMediaInFDB(dbName string, indexID string, mediaData []byte, rs *gjson.R
return
recordPath
,
errormdl
.
Wrap
(
"fail to add fdb index entry"
)
return
recordPath
,
errormdl
.
Wrap
(
"fail to add fdb index entry"
)
}
}
fileFpCache
.
Set
(
f
,
updatedInfileIndex
)
fileFpCache
.
Set
(
f
,
updatedInfileIndex
)
// json, err := createIndexJSON(index, rs)
// if err != nil {
// loggermdl.LogError(err)
// return recordPath, err
// }
rowID
,
err
:=
GenRowID
(
path
)
rowID
,
err
:=
GenRowID
(
path
)
if
err
!=
nil
{
if
err
!=
nil
{
loggermdl
.
LogError
(
err
)
loggermdl
.
LogError
(
err
)
...
@@ -1343,7 +1190,6 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r
...
@@ -1343,7 +1190,6 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r
resultStr
:=
"[]"
resultStr
:=
"[]"
errList
:=
[]
error
{}
errList
:=
[]
error
{}
updatedfps
:=
make
([]
*
os
.
File
,
0
)
updatedfps
:=
make
([]
*
os
.
File
,
0
)
// updatedCnt := 0
var
data
*
gjson
.
Result
var
data
*
gjson
.
Result
var
err
error
var
err
error
lazyEnable
:=
bucket
.
EnableLazy
lazyEnable
:=
bucket
.
EnableLazy
...
@@ -1363,17 +1209,8 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r
...
@@ -1363,17 +1209,8 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r
continue
continue
}
}
}
}
// filePath := filepath.Join(strings.TrimPrefix(fp.Name(), fdb.DBPath))
// path := filepath.Join(fdb.DBPath, filePath)
// if filePath != "" {
updatedfps
=
append
(
updatedfps
,
fp
)
updatedfps
=
append
(
updatedfps
,
fp
)
// filesUpdated = append(filesUpdated, filePath)
// filesUpdated[updatedCnt] = filePath
// updatedCnt++
resultStr
,
_
=
sjson
.
Set
(
resultStr
,
"-1"
,
data
.
Value
())
resultStr
,
_
=
sjson
.
Set
(
resultStr
,
"-1"
,
data
.
Value
())
// } else {
// loggermdl.LogError("invalid path for update - ", filePath)
// }
}
}
result
:=
gjson
.
Parse
(
resultStr
)
result
:=
gjson
.
Parse
(
resultStr
)
return
&
result
,
updatedfps
,
errList
return
&
result
,
updatedfps
,
errList
...
@@ -1408,7 +1245,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM
...
@@ -1408,7 +1245,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM
finalResultArray
:=
[]
gjson
.
Result
{}
finalResultArray
:=
[]
gjson
.
Result
{}
errList
:=
[]
error
{}
errList
:=
[]
error
{}
updatedfpInfileIndexMap
:=
make
(
map
[
*
os
.
File
]
*
gjson
.
Result
)
updatedfpInfileIndexMap
:=
make
(
map
[
*
os
.
File
]
*
gjson
.
Result
)
// dataUpdatedAtFilePaths := []string{}
secParams
:=
securitymdl
.
FDBSecParams
{
EnableSecurity
:
fdb
.
EnableSecurity
,
EnableCompression
:
fdb
.
EnableCompression
}
secParams
:=
securitymdl
.
FDBSecParams
{
EnableSecurity
:
fdb
.
EnableSecurity
,
EnableCompression
:
fdb
.
EnableCompression
}
for
fp
,
infileIndex
:=
range
fpInfileIndexMap
{
for
fp
,
infileIndex
:=
range
fpInfileIndexMap
{
...
@@ -1420,9 +1256,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM
...
@@ -1420,9 +1256,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM
continue
continue
}
}
updatedfpInfileIndexMap
[
fp
]
=
updatedInfileIndex
updatedfpInfileIndexMap
[
fp
]
=
updatedInfileIndex
// dataUpdatedAtFilePaths = append(dataUpdatedAtFilePaths, path)
// dataUpdatedAtFilePaths[dataUpdatedAtFilePathsCnt] = path
// dataUpdatedAtFilePathsCnt += len(resultArray.Array())
finalResultArray
=
append
(
finalResultArray
,
resultArray
.
Array
()
...
)
finalResultArray
=
append
(
finalResultArray
,
resultArray
.
Array
()
...
)
}
}
...
@@ -1533,8 +1366,6 @@ func AddIndexEntriesInFile(indexFilePath string, entries map[string]string) erro
...
@@ -1533,8 +1366,6 @@ func AddIndexEntriesInFile(indexFilePath string, entries map[string]string) erro
return
errormdl
.
CheckErr1
(
hashError
)
return
errormdl
.
CheckErr1
(
hashError
)
}
}
}
}
// dataByteToWriteRes := []byte(dataToStore)
//creating
return
filemdl
.
WriteFile
(
indexFilePath
,
dataByteToWriteRes
,
true
,
false
)
return
filemdl
.
WriteFile
(
indexFilePath
,
dataByteToWriteRes
,
true
,
false
)
}
}
...
@@ -1559,11 +1390,8 @@ func deleteDataFromNormalBucketUsingFp(index *Index, bucket *Bucket, fps []*os.F
...
@@ -1559,11 +1390,8 @@ func deleteDataFromNormalBucketUsingFp(index *Index, bucket *Bucket, fps []*os.F
errList
=
append
(
errList
,
errormdl
.
Wrap
(
"unable to delete file from lazy cache : "
+
err
.
Error
()))
errList
=
append
(
errList
,
errormdl
.
Wrap
(
"unable to delete file from lazy cache : "
+
err
.
Error
()))
}
}
}
}
// filePath := filepath.Join(fdbPath, path)
// delete file
// delete file
if
filemdl
.
FileAvailabilityCheck
(
fp
.
Name
())
{
if
filemdl
.
FileAvailabilityCheck
(
fp
.
Name
())
{
// path := fp.Name()
err
:=
fp
.
Truncate
(
0
)
err
:=
fp
.
Truncate
(
0
)
if
err
!=
nil
{
if
err
!=
nil
{
errList
=
append
(
errList
,
errormdl
.
Wrap
(
"unable to delete file : "
+
err
.
Error
()))
errList
=
append
(
errList
,
errormdl
.
Wrap
(
"unable to delete file : "
+
err
.
Error
()))
...
@@ -1819,8 +1647,6 @@ func SaveDataInFDB(dbName, indexID string, rs *gjson.Result) error {
...
@@ -1819,8 +1647,6 @@ func SaveDataInFDB(dbName, indexID string, rs *gjson.Result) error {
return
err
return
err
}
}
updatedJSONObj
:=
gjson
.
Parse
(
updatedJSON
)
updatedJSONObj
:=
gjson
.
Parse
(
updatedJSON
)
loggermdl
.
LogError
(
"index key"
,
path
)
loggermdl
.
LogError
(
"updatedJSONObj"
,
updatedJSONObj
)
err
=
index
.
AddEntry
(
path
,
&
updatedJSONObj
)
err
=
index
.
AddEntry
(
path
,
&
updatedJSONObj
)
if
err
!=
nil
{
if
err
!=
nil
{
loggermdl
.
LogError
(
err
)
loggermdl
.
LogError
(
err
)
...
@@ -1966,8 +1792,6 @@ func UpdateDataInFDB(dbName, indexID string, rs *gjson.Result, query []string, i
...
@@ -1966,8 +1792,6 @@ func UpdateDataInFDB(dbName, indexID string, rs *gjson.Result, query []string, i
// DeleteDataFromFDB -
// DeleteDataFromFDB -
func
DeleteDataFromFDB
(
dbName
string
,
indexID
string
,
rs
*
gjson
.
Result
,
queries
[]
string
,
infileIndexQueries
[]
string
)
(
recordsDeletedCnt
int
,
errList
[]
error
)
{
func
DeleteDataFromFDB
(
dbName
string
,
indexID
string
,
rs
*
gjson
.
Result
,
queries
[]
string
,
infileIndexQueries
[]
string
)
(
recordsDeletedCnt
int
,
errList
[]
error
)
{
// errList := make([]error, 0)
fdb
,
err
:=
GetFDBInstance
(
dbName
)
fdb
,
err
:=
GetFDBInstance
(
dbName
)
if
err
!=
nil
{
if
err
!=
nil
{
loggermdl
.
LogError
(
"fdb instance not found for: "
,
dbName
)
loggermdl
.
LogError
(
"fdb instance not found for: "
,
dbName
)
...
@@ -1984,7 +1808,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
...
@@ -1984,7 +1808,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
loggermdl
.
LogError
(
err
)
loggermdl
.
LogError
(
err
)
return
recordsDeletedCnt
,
[]
error
{
err
}
return
recordsDeletedCnt
,
[]
error
{
err
}
}
}
loggermdl
.
LogError
(
"indexKeyValMap"
,
indexKeyValMap
)
if
len
(
indexKeyValMap
)
==
0
{
if
len
(
indexKeyValMap
)
==
0
{
loggermdl
.
LogError
(
"no data found to delete"
)
loggermdl
.
LogError
(
"no data found to delete"
)
return
recordsDeletedCnt
,
[]
error
{
ErrNoDataFound
}
return
recordsDeletedCnt
,
[]
error
{
ErrNoDataFound
}
...
@@ -2007,7 +1830,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
...
@@ -2007,7 +1830,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
fpInfileIndexMap
[
rfile
.
file
]
=
rfile
.
InfileIndex
fpInfileIndexMap
[
rfile
.
file
]
=
rfile
.
InfileIndex
fps
=
append
(
fps
,
rfile
.
file
)
fps
=
append
(
fps
,
rfile
.
file
)
}
}
loggermdl
.
LogError
(
"update locking"
,
rfile
.
lock
)
rfile
.
lock
.
Lock
()
rfile
.
lock
.
Lock
()
rFiles
=
append
(
rFiles
,
rfile
)
rFiles
=
append
(
rFiles
,
rfile
)
...
@@ -2015,10 +1837,7 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
...
@@ -2015,10 +1837,7 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
defer
func
()
{
defer
func
()
{
for
i
:=
range
rFiles
{
for
i
:=
range
rFiles
{
loggermdl
.
LogError
(
"rfiles"
,
rFiles
[
i
]
.
lock
)
// loggermdl.LogError("update unlocking", rFiles[i].lock)
rFiles
[
i
]
.
lock
.
Unlock
()
rFiles
[
i
]
.
lock
.
Unlock
()
loggermdl
.
LogError
(
"after"
)
}
}
}()
}()
bucketID
:=
index
.
BucketSequence
[
len
(
index
.
BucketSequence
)
-
1
]
bucketID
:=
index
.
BucketSequence
[
len
(
index
.
BucketSequence
)
-
1
]
...
@@ -2079,8 +1898,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
...
@@ -2079,8 +1898,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
loggermdl
.
LogError
(
"Bucket not found: "
+
bucketID
)
loggermdl
.
LogError
(
"Bucket not found: "
+
bucketID
)
return
nil
,
errormdl
.
Wrap
(
"Bucket not found: "
+
bucketID
)
return
nil
,
errormdl
.
Wrap
(
"Bucket not found: "
+
bucketID
)
}
}
// filePaths := []string{}
// loggermdl.LogDebug(query)
indexKeyValueMap
,
err
:=
index
.
GetEntriesByQueries
(
queries
)
indexKeyValueMap
,
err
:=
index
.
GetEntriesByQueries
(
queries
)
if
err
!=
nil
{
if
err
!=
nil
{
loggermdl
.
LogError
(
err
)
loggermdl
.
LogError
(
err
)
...
@@ -2097,14 +1914,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
...
@@ -2097,14 +1914,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
for
filePath
:=
range
indexKeyValueMap
{
for
filePath
:=
range
indexKeyValueMap
{
filePath
=
filepath
.
Join
(
fdb
.
DBPath
,
filePath
)
filePath
=
filepath
.
Join
(
fdb
.
DBPath
,
filePath
)
loggermdl
.
LogError
(
"fileFpCache filePath"
,
filePath
)
rfile
,
err
:=
fileFpCache
.
Get
(
filePath
)
rfile
,
err
:=
fileFpCache
.
Get
(
filePath
)
if
err
!=
nil
{
if
err
!=
nil
{
fp
,
err
:=
openFile
(
filePath
)
fp
,
err
:=
openFile
(
filePath
)
if
err
!=
nil
{
if
err
!=
nil
{
continue
continue
}
}
// fileFpCache.Set(fp, nil)
fps
=
append
(
fps
,
fp
)
fps
=
append
(
fps
,
fp
)
fpInfileIndexMap
[
fp
]
=
nil
fpInfileIndexMap
[
fp
]
=
nil
rfile
=
fileFpCache
.
Set
(
fp
,
nil
)
rfile
=
fileFpCache
.
Set
(
fp
,
nil
)
...
@@ -2112,17 +1927,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
...
@@ -2112,17 +1927,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
fpInfileIndexMap
[
rfile
.
file
]
=
rfile
.
InfileIndex
fpInfileIndexMap
[
rfile
.
file
]
=
rfile
.
InfileIndex
fps
=
append
(
fps
,
rfile
.
file
)
fps
=
append
(
fps
,
rfile
.
file
)
}
}
loggermdl
.
LogError
(
"locking rFiles[i]"
,
rfile
.
file
)
loggermdl
.
LogError
(
"locking lock"
,
rfile
.
lock
)
rfile
.
lock
.
Lock
()
rfile
.
lock
.
Lock
()
loggermdl
.
LogError
(
"acquire lock"
,
rfile
.
lock
)
rFiles
=
append
(
rFiles
,
rfile
)
rFiles
=
append
(
rFiles
,
rfile
)
}
}
defer
func
()
{
defer
func
()
{
for
i
:=
range
rFiles
{
for
i
:=
range
rFiles
{
loggermdl
.
LogError
(
"unlocking rFiles[i]"
,
rFiles
[
i
]
.
file
)
loggermdl
.
LogError
(
"unlocking lock"
,
rFiles
[
i
]
.
lock
)
rFiles
[
i
]
.
lock
.
Unlock
()
rFiles
[
i
]
.
lock
.
Unlock
()
}
}
}()
}()
...
@@ -2154,7 +1964,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
...
@@ -2154,7 +1964,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
func
readDataFromPackFiles
(
filepathInfileIndexMap
map
[
*
os
.
File
]
*
gjson
.
Result
,
infileIndexQuery
[]
string
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
(
string
,
map
[
*
os
.
File
]
*
gjson
.
Result
,
error
)
{
func
readDataFromPackFiles
(
filepathInfileIndexMap
map
[
*
os
.
File
]
*
gjson
.
Result
,
infileIndexQuery
[]
string
,
rs
*
gjson
.
Result
,
secParams
securitymdl
.
FDBSecParams
)
(
string
,
map
[
*
os
.
File
]
*
gjson
.
Result
,
error
)
{
resultArray
:=
"[]"
resultArray
:=
"[]"
// filepathInfileIndexMap := make(map[string]*gjson.Result)
for
f
,
infileIndexData
:=
range
filepathInfileIndexMap
{
for
f
,
infileIndexData
:=
range
filepathInfileIndexMap
{
result
,
updatedInfileIndexData
,
err
:=
filepack
.
GetDataFromPackFileUsingFp
(
f
,
infileIndexData
,
infileIndexQuery
,
rs
,
secParams
)
result
,
updatedInfileIndexData
,
err
:=
filepack
.
GetDataFromPackFileUsingFp
(
f
,
infileIndexData
,
infileIndexQuery
,
rs
,
secParams
)
if
err
!=
nil
{
if
err
!=
nil
{
...
@@ -2192,7 +2001,6 @@ func addMigrationReplaceConfig(targetBasePath string, secParams securitymdl.FDBS
...
@@ -2192,7 +2001,6 @@ func addMigrationReplaceConfig(targetBasePath string, secParams securitymdl.FDBS
migartionConfigStr
,
_
:=
sjson
.
Set
(
""
,
MigrationTypeKeyword
,
MigrationTypeReplace
)
migartionConfigStr
,
_
:=
sjson
.
Set
(
""
,
MigrationTypeKeyword
,
MigrationTypeReplace
)
rs
:=
gjson
.
Parse
(
migartionConfigStr
)
rs
:=
gjson
.
Parse
(
migartionConfigStr
)
bucket
:=
Bucket
{}
bucket
:=
Bucket
{}
// err = saveDataInNormalBucketUsingFp(fp, rs)
return
saveDataInNormalBucketUsingFp
(
fp
,
&
bucket
,
&
rs
,
secParams
)
return
saveDataInNormalBucketUsingFp
(
fp
,
&
bucket
,
&
rs
,
secParams
)
}
}
...
@@ -2208,7 +2016,6 @@ func addMigrationUpdateConfig(targetBasePath string, secParams securitymdl.FDBSe
...
@@ -2208,7 +2016,6 @@ func addMigrationUpdateConfig(targetBasePath string, secParams securitymdl.FDBSe
migartionConfigStr
,
_
:=
sjson
.
Set
(
""
,
MigrationTypeKeyword
,
MigrationTypeUpdate
)
migartionConfigStr
,
_
:=
sjson
.
Set
(
""
,
MigrationTypeKeyword
,
MigrationTypeUpdate
)
rs
:=
gjson
.
Parse
(
migartionConfigStr
)
rs
:=
gjson
.
Parse
(
migartionConfigStr
)
bucket
:=
Bucket
{}
bucket
:=
Bucket
{}
// err = saveDataInNormalBucketUsingFp(fp, rs)
return
saveDataInNormalBucketUsingFp
(
fp
,
&
bucket
,
&
rs
,
secParams
)
return
saveDataInNormalBucketUsingFp
(
fp
,
&
bucket
,
&
rs
,
secParams
)
}
}
...
@@ -2255,8 +2062,6 @@ func (z ZipExporter) DataExport() (err error) {
...
@@ -2255,8 +2062,6 @@ func (z ZipExporter) DataExport() (err error) {
sourcePath
:=
""
sourcePath
:=
""
timeStamp
:=
time
.
Now
()
.
Nanosecond
()
timeStamp
:=
time
.
Now
()
.
Nanosecond
()
targetBasePath
:=
filepath
.
Join
(
filemdl
.
TempDir
,
strconv
.
Itoa
(
timeStamp
))
targetBasePath
:=
filepath
.
Join
(
filemdl
.
TempDir
,
strconv
.
Itoa
(
timeStamp
))
// loggermdl.LogDebug("targetBasePath", targetBasePath)
// filepathsToExport := []string{}
filteredKeyValMap
,
err
:=
index
.
GetEntriesByQueries
(
z
.
Queries
)
filteredKeyValMap
,
err
:=
index
.
GetEntriesByQueries
(
z
.
Queries
)
if
err
!=
nil
{
if
err
!=
nil
{
return
err
return
err
...
@@ -2279,8 +2084,6 @@ func (z ZipExporter) DataExport() (err error) {
...
@@ -2279,8 +2084,6 @@ func (z ZipExporter) DataExport() (err error) {
}
}
// copy index file
// copy index file
// sourcePath = filepath.Join(fdb.DBPath, INDEXFOLDER, indexID)
// _, err = filemdl.CopyFile(sourcePath, targetPath, true)
targetPath
:=
filepath
.
Join
(
targetBasePath
,
INDEXFOLDER
,
z
.
IndexID
)
targetPath
:=
filepath
.
Join
(
targetBasePath
,
INDEXFOLDER
,
z
.
IndexID
)
err
=
AddIndexEntriesInFile
(
targetPath
,
filteredKeyValMap
)
err
=
AddIndexEntriesInFile
(
targetPath
,
filteredKeyValMap
)
if
err
!=
nil
{
if
err
!=
nil
{
...
@@ -2476,14 +2279,7 @@ func ImportIndexEntries(indexFilePath string, fdb *FDB, indexID string) error {
...
@@ -2476,14 +2279,7 @@ func ImportIndexEntries(indexFilePath string, fdb *FDB, indexID string) error {
}
}
// update index file by reading all data and updating index file
// update index file by reading all data and updating index file
err
=
index
.
AddEntries
(
indexDataMap
)
return
index
.
AddEntries
(
indexDataMap
)
// for key, val := range indexDataMap {
// err := index.AddEntry(key, val)
// }
// if err != nil {
// return err
// }
return
err
}
}
// FileImporter is a DataImporter
// FileImporter is a DataImporter
...
...
This diff is collapsed.
Click to expand it.
dalmdl/corefdb/indexStore.go
+
0
−
2
View file @
342e46f8
...
@@ -4,7 +4,6 @@ import (
...
@@ -4,7 +4,6 @@ import (
"path/filepath"
"path/filepath"
"strings"
"strings"
"corelab.mkcl.org/MKCLOS/coredevelopmentplatform/corepkgv2/loggermdl"
"corelab.mkcl.org/MKCLOS/coredevelopmentplatform/corepkgv2/securitymdl"
"corelab.mkcl.org/MKCLOS/coredevelopmentplatform/corepkgv2/securitymdl"
"github.com/tidwall/buntdb"
"github.com/tidwall/buntdb"
...
@@ -100,7 +99,6 @@ func (i *indexStore) GetManyByQuery(queries []string) (map[string]string, error)
...
@@ -100,7 +99,6 @@ func (i *indexStore) GetManyByQuery(queries []string) (map[string]string, error)
entryMap
:=
make
(
map
[
string
]
string
,
0
)
entryMap
:=
make
(
map
[
string
]
string
,
0
)
err
:=
i
.
store
.
View
(
func
(
tx
*
buntdb
.
Tx
)
error
{
err
:=
i
.
store
.
View
(
func
(
tx
*
buntdb
.
Tx
)
error
{
return
tx
.
Ascend
(
""
,
func
(
key
,
value
string
)
bool
{
return
tx
.
Ascend
(
""
,
func
(
key
,
value
string
)
bool
{
loggermdl
.
LogError
(
"key"
,
key
,
"val"
,
value
)
rsJSON
:=
gjson
.
Parse
(
"["
+
value
+
"]"
)
rsJSON
:=
gjson
.
Parse
(
"["
+
value
+
"]"
)
for
i
:=
0
;
i
<
len
(
queries
);
i
++
{
for
i
:=
0
;
i
<
len
(
queries
);
i
++
{
rsJSON
=
rsJSON
.
Get
(
queries
[
i
]
+
"#"
)
rsJSON
=
rsJSON
.
Get
(
queries
[
i
]
+
"#"
)
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment
Menu
Explore
Projects
Groups
Topics
Snippets