Commit 342e46f8 authored by Vivek Naik's avatar Vivek Naik
Browse files

removed commented code

parent 36f3d0b6
2 merge requests!169Mep release 060320,!167fdb fixes
Showing with 1 addition and 207 deletions
...@@ -145,11 +145,6 @@ func (fdb *FDB) NewBucket(bucket, parentBucket *Bucket) *Bucket { ...@@ -145,11 +145,6 @@ func (fdb *FDB) NewBucket(bucket, parentBucket *Bucket) *Bucket {
return bucket return bucket
} }
// // SetIsBucketPackType - SetIsBucketPackType
// func (bucket *Bucket) SetIsBucketPackType(isPackType bool) {
// bucket.IsSinglePackType = isPackType
// }
// SetBucketType - // SetBucketType -
func (bucket *Bucket) SetBucketType(bucketType string) { func (bucket *Bucket) SetBucketType(bucketType string) {
bucket.BucketType = bucketType bucket.BucketType = bucketType
...@@ -318,35 +313,6 @@ func AppendDataInLazyObjectInCache(bucketID string, data gjson.Result, filePath ...@@ -318,35 +313,6 @@ func AppendDataInLazyObjectInCache(bucketID string, data gjson.Result, filePath
return nil return nil
} }
// // restoreFilesFromFilePack - restoreFilesFromFilePack
// func restoreFilesFromFilePack(fileName string, fileData []byte) error {
// if len(fileData) > 0 {
// err := filemdl.WriteFile(fileName, fileData, true, false)
// if errormdl.CheckErr(err) != nil {
// loggermdl.LogError(err)
// return errormdl.CheckErr(err)
// }
// }
// return nil
// }
// func readNormalFile(path string) ([]byte, error) {
// ba, err := filemdl.ReadFile(path)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// if isSecurityEnabled {
// _, fileName := filepath.Split(path)
// ba, err = decryptData(ba, fileName)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// }
// return ba, nil
// }
func readNormalFileUsingFp(fp *os.File, rs *gjson.Result, secParams securitymdl.FDBSecParams) ([]byte, error) { func readNormalFileUsingFp(fp *os.File, rs *gjson.Result, secParams securitymdl.FDBSecParams) ([]byte, error) {
ba, err := filemdl.ReadFileUsingFp(fp) ba, err := filemdl.ReadFileUsingFp(fp)
if err != nil { if err != nil {
...@@ -374,37 +340,6 @@ func readNormalFileUsingFp(fp *os.File, rs *gjson.Result, secParams securitymdl. ...@@ -374,37 +340,6 @@ func readNormalFileUsingFp(fp *os.File, rs *gjson.Result, secParams securitymdl.
return ba, nil return ba, nil
} }
// // readNormalFiles - reads file and returns file data
// func (fdb *FDB) readNormalFiles(path []string) (*gjson.Result, error) {
// sb := strings.Builder{}
// sb.WriteString("[")
// for i := 0; i < len(path); i++ {
// updatedPath := filepath.Join(fdb.DBPath, path[i])
// ba := []byte{'{', '}'}
// var err error
// if filemdl.FileAvailabilityCheck(updatedPath) {
// ba, err = readNormalFile(updatedPath)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// // loggermdl.LogError("File data for index = ", i, " - ", string(ba))
// _, err = sb.Write(ba)
// if err != nil {
// loggermdl.LogError(err)
// return nil, err
// }
// sb.WriteString(",")
// }
// }
// sb.WriteString("]")
// finalResult := strings.Replace(sb.String(), ",]", "]", 1)
// rs := gjson.Parse(finalResult)
// return &rs, nil
// }
// readNormalFilesUsingFp - reads file and returns file data
func readNormalFilesUsingFp(fps []*os.File, bucket *Bucket, rs *gjson.Result, secParams securitymdl.FDBSecParams) (*gjson.Result, error) { func readNormalFilesUsingFp(fps []*os.File, bucket *Bucket, rs *gjson.Result, secParams securitymdl.FDBSecParams) (*gjson.Result, error) {
sb := strings.Builder{} sb := strings.Builder{}
...@@ -428,7 +363,6 @@ func readNormalFilesUsingFp(fps []*os.File, bucket *Bucket, rs *gjson.Result, se ...@@ -428,7 +363,6 @@ func readNormalFilesUsingFp(fps []*os.File, bucket *Bucket, rs *gjson.Result, se
loggermdl.LogError(err) loggermdl.LogError(err)
return nil, err return nil, err
} }
// loggermdl.LogError("File data for index = ", i, " - ", string(ba))
_, err = sb.Write(ba) _, err = sb.Write(ba)
if err != nil { if err != nil {
loggermdl.LogError(err) loggermdl.LogError(err)
...@@ -716,28 +650,6 @@ func deleteDataFromLazyBucket(bucket *Bucket, fp *os.File) error { ...@@ -716,28 +650,6 @@ func deleteDataFromLazyBucket(bucket *Bucket, fp *os.File) error {
return nil return nil
} }
// func saveDataInNormalBucketUsingFpNew(f *os.File, bucket *Bucket, rs *gjson.Result) error {
// dataBytes := []byte(rs.String())
// var err error
// if bucket.EnableLazy {
// return saveDatInLazyBucket(bucket, f, rs)
// }
// if isSecurityEnabled {
// _, fileName := filepath.Split(f.Name())
// dataBytes, err = encryptData(dataBytes, fileName)
// if err != nil {
// loggermdl.LogError(err)
// return err
// }
// }
// err = filemdl.WriteFileUsingFp(f, dataBytes, true, false)
// if errormdl.CheckErr(err) != nil {
// loggermdl.LogError(err)
// return errormdl.CheckErr(err)
// }
// return nil
// }
func saveDataInNormalBucketUsingFp(f *os.File, bucket *Bucket, rs *gjson.Result, secParams securitymdl.FDBSecParams) error { func saveDataInNormalBucketUsingFp(f *os.File, bucket *Bucket, rs *gjson.Result, secParams securitymdl.FDBSecParams) error {
dataBytes := []byte(rs.String()) dataBytes := []byte(rs.String())
var err error var err error
...@@ -774,66 +686,6 @@ func saveDataInNormalBucketUsingFp(f *os.File, bucket *Bucket, rs *gjson.Result, ...@@ -774,66 +686,6 @@ func saveDataInNormalBucketUsingFp(f *os.File, bucket *Bucket, rs *gjson.Result,
return nil return nil
} }
// func saveDataInNormalBucket(filePath string, rs *gjson.Result) error {
// dataBytes := []byte(rs.String())
// var err error
// if isSecurityEnabled {
// _, fileName := filepath.Split(filePath)
// dataBytes, err = encryptData(dataBytes, fileName)
// if err != nil {
// loggermdl.LogError(err)
// return err
// }
// }
// err = filemdl.WriteFile(filePath, dataBytes, true, false)
// if errormdl.CheckErr(err) != nil {
// loggermdl.LogError(err)
// return errormdl.CheckErr(err)
// }
// return nil
// }
// func saveDataInPackBucket(fdb *FDB, filePath string, index *Index, rs *gjson.Result) error {
// // check is file locked
// // if yes
// // wait
// // else
// // /* */update filestatus to fileStatusUpdatingData
// // compress and encrypt data
// // calculate filehash
// //save data in file & getoffset and len
// //update footer address
// // update filestatuse to fileStatusUpdatingIndex
// // update index
// // update filestatus to fileStatusReady
// bucketID := index.BucketSequence[len(index.BucketSequence)-1]
// bucket := fdb.buckets[bucketID]
// fileType := rs.Get("fileType").String()
// if len(fileType) == 0 {
// return errormdl.Wrap("please specify fileType")
// }
// isFilePresent := filemdl.FileAvailabilityCheck(filePath)
// infileIndex, ok := bucket.InFileIndexMap[fileType]
// if !ok {
// loggermdl.LogError("infileIndex for specified fileType not found: ", fileType)
// return errormdl.Wrap("infileIndex for specified fileType not found: " + fileType)
// }
// err := filepack.AddFileInPackFile(filePath, infileIndex.IndexFields, rs)
// if err != nil {
// loggermdl.LogError("fail to add fdb index entry in file: ", err)
// return errormdl.Wrap("fail to write data in file")
// }
// // for adding fdb index data
// if !isFilePresent {
// err = addFDBIndexEntryFile(filePath, bucket, index.IndexFields, rs)
// }
// return err
// }
func saveDataInPackBucketUsingFp(fdb *FDB, f *os.File, index *Index, infileIndexData, rs *gjson.Result) (*gjson.Result, error) { func saveDataInPackBucketUsingFp(fdb *FDB, f *os.File, index *Index, infileIndexData, rs *gjson.Result) (*gjson.Result, error) {
// check is file locked // check is file locked
// if yes // if yes
...@@ -944,11 +796,6 @@ func SaveMediaInFDB(dbName string, indexID string, mediaData []byte, rs *gjson.R ...@@ -944,11 +796,6 @@ func SaveMediaInFDB(dbName string, indexID string, mediaData []byte, rs *gjson.R
return recordPath, errormdl.Wrap("fail to add fdb index entry") return recordPath, errormdl.Wrap("fail to add fdb index entry")
} }
fileFpCache.Set(f, updatedInfileIndex) fileFpCache.Set(f, updatedInfileIndex)
// json, err := createIndexJSON(index, rs)
// if err != nil {
// loggermdl.LogError(err)
// return recordPath, err
// }
rowID, err := GenRowID(path) rowID, err := GenRowID(path)
if err != nil { if err != nil {
loggermdl.LogError(err) loggermdl.LogError(err)
...@@ -1343,7 +1190,6 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r ...@@ -1343,7 +1190,6 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r
resultStr := "[]" resultStr := "[]"
errList := []error{} errList := []error{}
updatedfps := make([]*os.File, 0) updatedfps := make([]*os.File, 0)
// updatedCnt := 0
var data *gjson.Result var data *gjson.Result
var err error var err error
lazyEnable := bucket.EnableLazy lazyEnable := bucket.EnableLazy
...@@ -1363,17 +1209,8 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r ...@@ -1363,17 +1209,8 @@ func updateDataInNormalBucketUsingFp(fdb *FDB, bucket *Bucket, fps []*os.File, r
continue continue
} }
} }
// filePath := filepath.Join(strings.TrimPrefix(fp.Name(), fdb.DBPath))
// path := filepath.Join(fdb.DBPath, filePath)
// if filePath != "" {
updatedfps = append(updatedfps, fp) updatedfps = append(updatedfps, fp)
// filesUpdated = append(filesUpdated, filePath)
// filesUpdated[updatedCnt] = filePath
// updatedCnt++
resultStr, _ = sjson.Set(resultStr, "-1", data.Value()) resultStr, _ = sjson.Set(resultStr, "-1", data.Value())
// } else {
// loggermdl.LogError("invalid path for update - ", filePath)
// }
} }
result := gjson.Parse(resultStr) result := gjson.Parse(resultStr)
return &result, updatedfps, errList return &result, updatedfps, errList
...@@ -1408,7 +1245,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM ...@@ -1408,7 +1245,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM
finalResultArray := []gjson.Result{} finalResultArray := []gjson.Result{}
errList := []error{} errList := []error{}
updatedfpInfileIndexMap := make(map[*os.File]*gjson.Result) updatedfpInfileIndexMap := make(map[*os.File]*gjson.Result)
// dataUpdatedAtFilePaths := []string{}
secParams := securitymdl.FDBSecParams{EnableSecurity: fdb.EnableSecurity, EnableCompression: fdb.EnableCompression} secParams := securitymdl.FDBSecParams{EnableSecurity: fdb.EnableSecurity, EnableCompression: fdb.EnableCompression}
for fp, infileIndex := range fpInfileIndexMap { for fp, infileIndex := range fpInfileIndexMap {
...@@ -1420,9 +1256,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM ...@@ -1420,9 +1256,6 @@ func updateDataInFileIndexBucketUsingFp(fdb *FDB, bucket *Bucket, fpInfileIndexM
continue continue
} }
updatedfpInfileIndexMap[fp] = updatedInfileIndex updatedfpInfileIndexMap[fp] = updatedInfileIndex
// dataUpdatedAtFilePaths = append(dataUpdatedAtFilePaths, path)
// dataUpdatedAtFilePaths[dataUpdatedAtFilePathsCnt] = path
// dataUpdatedAtFilePathsCnt += len(resultArray.Array())
finalResultArray = append(finalResultArray, resultArray.Array()...) finalResultArray = append(finalResultArray, resultArray.Array()...)
} }
...@@ -1533,8 +1366,6 @@ func AddIndexEntriesInFile(indexFilePath string, entries map[string]string) erro ...@@ -1533,8 +1366,6 @@ func AddIndexEntriesInFile(indexFilePath string, entries map[string]string) erro
return errormdl.CheckErr1(hashError) return errormdl.CheckErr1(hashError)
} }
} }
// dataByteToWriteRes := []byte(dataToStore)
//creating
return filemdl.WriteFile(indexFilePath, dataByteToWriteRes, true, false) return filemdl.WriteFile(indexFilePath, dataByteToWriteRes, true, false)
} }
...@@ -1559,11 +1390,8 @@ func deleteDataFromNormalBucketUsingFp(index *Index, bucket *Bucket, fps []*os.F ...@@ -1559,11 +1390,8 @@ func deleteDataFromNormalBucketUsingFp(index *Index, bucket *Bucket, fps []*os.F
errList = append(errList, errormdl.Wrap("unable to delete file from lazy cache : "+err.Error())) errList = append(errList, errormdl.Wrap("unable to delete file from lazy cache : "+err.Error()))
} }
} }
// filePath := filepath.Join(fdbPath, path)
// delete file // delete file
if filemdl.FileAvailabilityCheck(fp.Name()) { if filemdl.FileAvailabilityCheck(fp.Name()) {
// path := fp.Name()
err := fp.Truncate(0) err := fp.Truncate(0)
if err != nil { if err != nil {
errList = append(errList, errormdl.Wrap("unable to delete file : "+err.Error())) errList = append(errList, errormdl.Wrap("unable to delete file : "+err.Error()))
...@@ -1819,8 +1647,6 @@ func SaveDataInFDB(dbName, indexID string, rs *gjson.Result) error { ...@@ -1819,8 +1647,6 @@ func SaveDataInFDB(dbName, indexID string, rs *gjson.Result) error {
return err return err
} }
updatedJSONObj := gjson.Parse(updatedJSON) updatedJSONObj := gjson.Parse(updatedJSON)
loggermdl.LogError("index key", path)
loggermdl.LogError("updatedJSONObj", updatedJSONObj)
err = index.AddEntry(path, &updatedJSONObj) err = index.AddEntry(path, &updatedJSONObj)
if err != nil { if err != nil {
loggermdl.LogError(err) loggermdl.LogError(err)
...@@ -1966,8 +1792,6 @@ func UpdateDataInFDB(dbName, indexID string, rs *gjson.Result, query []string, i ...@@ -1966,8 +1792,6 @@ func UpdateDataInFDB(dbName, indexID string, rs *gjson.Result, query []string, i
// DeleteDataFromFDB - // DeleteDataFromFDB -
func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries []string, infileIndexQueries []string) (recordsDeletedCnt int, errList []error) { func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries []string, infileIndexQueries []string) (recordsDeletedCnt int, errList []error) {
// errList := make([]error, 0)
fdb, err := GetFDBInstance(dbName) fdb, err := GetFDBInstance(dbName)
if err != nil { if err != nil {
loggermdl.LogError("fdb instance not found for: ", dbName) loggermdl.LogError("fdb instance not found for: ", dbName)
...@@ -1984,7 +1808,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries ...@@ -1984,7 +1808,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
loggermdl.LogError(err) loggermdl.LogError(err)
return recordsDeletedCnt, []error{err} return recordsDeletedCnt, []error{err}
} }
loggermdl.LogError("indexKeyValMap", indexKeyValMap)
if len(indexKeyValMap) == 0 { if len(indexKeyValMap) == 0 {
loggermdl.LogError("no data found to delete") loggermdl.LogError("no data found to delete")
return recordsDeletedCnt, []error{ErrNoDataFound} return recordsDeletedCnt, []error{ErrNoDataFound}
...@@ -2007,7 +1830,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries ...@@ -2007,7 +1830,6 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
fpInfileIndexMap[rfile.file] = rfile.InfileIndex fpInfileIndexMap[rfile.file] = rfile.InfileIndex
fps = append(fps, rfile.file) fps = append(fps, rfile.file)
} }
loggermdl.LogError("update locking", rfile.lock)
rfile.lock.Lock() rfile.lock.Lock()
rFiles = append(rFiles, rfile) rFiles = append(rFiles, rfile)
...@@ -2015,10 +1837,7 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries ...@@ -2015,10 +1837,7 @@ func DeleteDataFromFDB(dbName string, indexID string, rs *gjson.Result, queries
defer func() { defer func() {
for i := range rFiles { for i := range rFiles {
loggermdl.LogError("rfiles", rFiles[i].lock)
// loggermdl.LogError("update unlocking", rFiles[i].lock)
rFiles[i].lock.Unlock() rFiles[i].lock.Unlock()
loggermdl.LogError("after")
} }
}() }()
bucketID := index.BucketSequence[len(index.BucketSequence)-1] bucketID := index.BucketSequence[len(index.BucketSequence)-1]
...@@ -2079,8 +1898,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string, ...@@ -2079,8 +1898,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
loggermdl.LogError("Bucket not found: " + bucketID) loggermdl.LogError("Bucket not found: " + bucketID)
return nil, errormdl.Wrap("Bucket not found: " + bucketID) return nil, errormdl.Wrap("Bucket not found: " + bucketID)
} }
// filePaths := []string{}
// loggermdl.LogDebug(query)
indexKeyValueMap, err := index.GetEntriesByQueries(queries) indexKeyValueMap, err := index.GetEntriesByQueries(queries)
if err != nil { if err != nil {
loggermdl.LogError(err) loggermdl.LogError(err)
...@@ -2097,14 +1914,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string, ...@@ -2097,14 +1914,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
for filePath := range indexKeyValueMap { for filePath := range indexKeyValueMap {
filePath = filepath.Join(fdb.DBPath, filePath) filePath = filepath.Join(fdb.DBPath, filePath)
loggermdl.LogError("fileFpCache filePath", filePath)
rfile, err := fileFpCache.Get(filePath) rfile, err := fileFpCache.Get(filePath)
if err != nil { if err != nil {
fp, err := openFile(filePath) fp, err := openFile(filePath)
if err != nil { if err != nil {
continue continue
} }
// fileFpCache.Set(fp, nil)
fps = append(fps, fp) fps = append(fps, fp)
fpInfileIndexMap[fp] = nil fpInfileIndexMap[fp] = nil
rfile = fileFpCache.Set(fp, nil) rfile = fileFpCache.Set(fp, nil)
...@@ -2112,17 +1927,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string, ...@@ -2112,17 +1927,12 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
fpInfileIndexMap[rfile.file] = rfile.InfileIndex fpInfileIndexMap[rfile.file] = rfile.InfileIndex
fps = append(fps, rfile.file) fps = append(fps, rfile.file)
} }
loggermdl.LogError("locking rFiles[i]", rfile.file)
loggermdl.LogError("locking lock", rfile.lock)
rfile.lock.Lock() rfile.lock.Lock()
loggermdl.LogError("acquire lock", rfile.lock)
rFiles = append(rFiles, rfile) rFiles = append(rFiles, rfile)
} }
defer func() { defer func() {
for i := range rFiles { for i := range rFiles {
loggermdl.LogError("unlocking rFiles[i]", rFiles[i].file)
loggermdl.LogError("unlocking lock", rFiles[i].lock)
rFiles[i].lock.Unlock() rFiles[i].lock.Unlock()
} }
}() }()
...@@ -2154,7 +1964,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string, ...@@ -2154,7 +1964,6 @@ func ReadDataFromFDB(dbName, indexID string, rs *gjson.Result, queries []string,
func readDataFromPackFiles(filepathInfileIndexMap map[*os.File]*gjson.Result, infileIndexQuery []string, rs *gjson.Result, secParams securitymdl.FDBSecParams) (string, map[*os.File]*gjson.Result, error) { func readDataFromPackFiles(filepathInfileIndexMap map[*os.File]*gjson.Result, infileIndexQuery []string, rs *gjson.Result, secParams securitymdl.FDBSecParams) (string, map[*os.File]*gjson.Result, error) {
resultArray := "[]" resultArray := "[]"
// filepathInfileIndexMap := make(map[string]*gjson.Result)
for f, infileIndexData := range filepathInfileIndexMap { for f, infileIndexData := range filepathInfileIndexMap {
result, updatedInfileIndexData, err := filepack.GetDataFromPackFileUsingFp(f, infileIndexData, infileIndexQuery, rs, secParams) result, updatedInfileIndexData, err := filepack.GetDataFromPackFileUsingFp(f, infileIndexData, infileIndexQuery, rs, secParams)
if err != nil { if err != nil {
...@@ -2192,7 +2001,6 @@ func addMigrationReplaceConfig(targetBasePath string, secParams securitymdl.FDBS ...@@ -2192,7 +2001,6 @@ func addMigrationReplaceConfig(targetBasePath string, secParams securitymdl.FDBS
migartionConfigStr, _ := sjson.Set("", MigrationTypeKeyword, MigrationTypeReplace) migartionConfigStr, _ := sjson.Set("", MigrationTypeKeyword, MigrationTypeReplace)
rs := gjson.Parse(migartionConfigStr) rs := gjson.Parse(migartionConfigStr)
bucket := Bucket{} bucket := Bucket{}
// err = saveDataInNormalBucketUsingFp(fp, rs)
return saveDataInNormalBucketUsingFp(fp, &bucket, &rs, secParams) return saveDataInNormalBucketUsingFp(fp, &bucket, &rs, secParams)
} }
...@@ -2208,7 +2016,6 @@ func addMigrationUpdateConfig(targetBasePath string, secParams securitymdl.FDBSe ...@@ -2208,7 +2016,6 @@ func addMigrationUpdateConfig(targetBasePath string, secParams securitymdl.FDBSe
migartionConfigStr, _ := sjson.Set("", MigrationTypeKeyword, MigrationTypeUpdate) migartionConfigStr, _ := sjson.Set("", MigrationTypeKeyword, MigrationTypeUpdate)
rs := gjson.Parse(migartionConfigStr) rs := gjson.Parse(migartionConfigStr)
bucket := Bucket{} bucket := Bucket{}
// err = saveDataInNormalBucketUsingFp(fp, rs)
return saveDataInNormalBucketUsingFp(fp, &bucket, &rs, secParams) return saveDataInNormalBucketUsingFp(fp, &bucket, &rs, secParams)
} }
...@@ -2255,8 +2062,6 @@ func (z ZipExporter) DataExport() (err error) { ...@@ -2255,8 +2062,6 @@ func (z ZipExporter) DataExport() (err error) {
sourcePath := "" sourcePath := ""
timeStamp := time.Now().Nanosecond() timeStamp := time.Now().Nanosecond()
targetBasePath := filepath.Join(filemdl.TempDir, strconv.Itoa(timeStamp)) targetBasePath := filepath.Join(filemdl.TempDir, strconv.Itoa(timeStamp))
// loggermdl.LogDebug("targetBasePath", targetBasePath)
// filepathsToExport := []string{}
filteredKeyValMap, err := index.GetEntriesByQueries(z.Queries) filteredKeyValMap, err := index.GetEntriesByQueries(z.Queries)
if err != nil { if err != nil {
return err return err
...@@ -2279,8 +2084,6 @@ func (z ZipExporter) DataExport() (err error) { ...@@ -2279,8 +2084,6 @@ func (z ZipExporter) DataExport() (err error) {
} }
// copy index file // copy index file
// sourcePath = filepath.Join(fdb.DBPath, INDEXFOLDER, indexID)
// _, err = filemdl.CopyFile(sourcePath, targetPath, true)
targetPath := filepath.Join(targetBasePath, INDEXFOLDER, z.IndexID) targetPath := filepath.Join(targetBasePath, INDEXFOLDER, z.IndexID)
err = AddIndexEntriesInFile(targetPath, filteredKeyValMap) err = AddIndexEntriesInFile(targetPath, filteredKeyValMap)
if err != nil { if err != nil {
...@@ -2476,14 +2279,7 @@ func ImportIndexEntries(indexFilePath string, fdb *FDB, indexID string) error { ...@@ -2476,14 +2279,7 @@ func ImportIndexEntries(indexFilePath string, fdb *FDB, indexID string) error {
} }
// update index file by reading all data and updating index file // update index file by reading all data and updating index file
err = index.AddEntries(indexDataMap) return index.AddEntries(indexDataMap)
// for key, val := range indexDataMap {
// err := index.AddEntry(key, val)
// }
// if err != nil {
// return err
// }
return err
} }
// FileImporter is a DataImporter // FileImporter is a DataImporter
......
...@@ -4,7 +4,6 @@ import ( ...@@ -4,7 +4,6 @@ import (
"path/filepath" "path/filepath"
"strings" "strings"
"corelab.mkcl.org/MKCLOS/coredevelopmentplatform/corepkgv2/loggermdl"
"corelab.mkcl.org/MKCLOS/coredevelopmentplatform/corepkgv2/securitymdl" "corelab.mkcl.org/MKCLOS/coredevelopmentplatform/corepkgv2/securitymdl"
"github.com/tidwall/buntdb" "github.com/tidwall/buntdb"
...@@ -100,7 +99,6 @@ func (i *indexStore) GetManyByQuery(queries []string) (map[string]string, error) ...@@ -100,7 +99,6 @@ func (i *indexStore) GetManyByQuery(queries []string) (map[string]string, error)
entryMap := make(map[string]string, 0) entryMap := make(map[string]string, 0)
err := i.store.View(func(tx *buntdb.Tx) error { err := i.store.View(func(tx *buntdb.Tx) error {
return tx.Ascend("", func(key, value string) bool { return tx.Ascend("", func(key, value string) bool {
loggermdl.LogError("key", key, "val", value)
rsJSON := gjson.Parse("[" + value + "]") rsJSON := gjson.Parse("[" + value + "]")
for i := 0; i < len(queries); i++ { for i := 0; i < len(queries); i++ {
rsJSON = rsJSON.Get(queries[i] + "#") rsJSON = rsJSON.Get(queries[i] + "#")
......
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment