Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fs insert one new #166

Open
wants to merge 16 commits into
base: client
Choose a base branch
from
Open

Conversation

sl5035
Copy link

@sl5035 sl5035 commented Mar 21, 2023

No description provided.

@sl5035
Copy link
Author

sl5035 commented Mar 21, 2023

Sorry I had to change many files in order to make the function work.

  • Anything related to ChainDB had to be commented out since I did not write ChainDB module yet. I'm considering removing it from pydatarecognition after I'm done with fsclient.
  • Also, what does db['blacklist'] and db['whitelist'] do? I deleted it to run fsclient for now.
  • I am still figuring out how to test invalid inputs.

@@ -15,7 +15,7 @@
from copy import deepcopy


OUTPUT_FAKE_DB = False # always turn it to false after you used it
OUTPUT_FAKE_DB = True # always turn it to false after you used it
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sorry I forgot to turn it to false.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please change to false and push

Copy link
Collaborator

@sbillinge sbillinge left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this looks good except we need a new push with the db output set to False

@@ -15,7 +15,7 @@
from copy import deepcopy


OUTPUT_FAKE_DB = False # always turn it to false after you used it
OUTPUT_FAKE_DB = True # always turn it to false after you used it
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this should be False

@@ -15,7 +15,7 @@
from copy import deepcopy


OUTPUT_FAKE_DB = False # always turn it to false after you used it
OUTPUT_FAKE_DB = True # always turn it to false after you used it
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please change to false and push

Copy link
Collaborator

@sbillinge sbillinge left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

pls see comments

@pytest.mark.skip("Not written")
def test_insert_one():
pass
test_insert_cif = [{'intensity': [], 'q': [], 'ttheta': [], 'wavelength': 0.111111, '_id': 'ts1129'},
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

let's think about all the things we want to test. Since insert_one is general (not just cifs in principle) we just need to validate that it is valid json or sthg?

def test_insert_one(make_db, make_bad_db, tc):
db_path = make_db
pydr_rc['databases'][0]['url'] = db_path
rc._update(pydr_rc) # TODO: Is there a way to update rc at a global scope so that we don't have to write this every time we run the test functions?
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

you may be able to build rc in conftest.py and import it?

@sl5035
Copy link
Author

sl5035 commented Mar 23, 2023

Hello Professor,
Could you please confirm test_insert_one and insert_one? If there is nothing to change I will move on to find one after this branch gets merged. Thank you!

Copy link
Collaborator

@sbillinge sbillinge left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sorry, I thought I sent these comments before...just finding them now as unsent.....

client.insert_one(dbname, collname, tc)

assert list(client.dbs[dbname][collname].values())[0] == tc # TODO: How to reformat
try:
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

remember, don't put logic in the test. This is passing because you handle it in the test, but it needs to pass because you handle it in the way you want in the function itself.

{'intensity': [], 'q': [], 'ttheta': [], 'wavelength': 0.111111, '_id': 'ts1129'}),
('bad_case_test', 'bad_case_test')]
@pytest.mark.parametrize('input, result', test_insert_cif)
def test_insert_one(make_db, rc, input, result):
client = FileSystemClient(rc)
client.open()

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

the code below seems to be testing connect_db, which should be tested in test_connect_db and not here. It is ok to use the function here to connect your db, but don't write tests for it here.


assert list(client.dbs[dbname][collname].values())[0] == tc # TODO: How to reformat
try:
client.insert_one(dbname, collname, input)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's maybe take a step back here and decide what behavior we want. I think we want insert_one to just insert a document successfully into the client database, but we don't want it to do any validation etc.. Do you agree? Then when we use it we may want the validation so perhaps it would be something like:

cif_doc = create_doc(cif_info)  # I made tihs up, we don't have this function or need it necessarily but it captures the behavior
cif_doc_bool = validate_cifness(cif_doc)
if cif_doc_bool:
    client.insert_one(db, coll, cif_doc)
else:
    something

I am not sure if this is exactly right, but something like that. This makes the functions more reusable because I could use client.insert_one() to insert something that is not a cif document, so it is more reusable.

In this case, I think what we need to test is that insert_one does insert the thing properly, so we need a file on the TempDir filesystem, then after the insert_one we need to read it and show that nothing has changed in there except that the new doc has been inserted correctly. If the input is invalid we need to make sure that the file remains unchanged. sthg like that?

@sbillinge
Copy link
Collaborator

sbillinge commented Mar 23, 2023 via email

@sl5035
Copy link
Author

sl5035 commented Mar 23, 2023

sorry, I thought I sent these comments before...just finding them now as unsent.....

All good! I see them now. I'll fix these asap.

@sl5035
Copy link
Author

sl5035 commented Mar 23, 2023

Using a bad test case, the function does not insert into the database so the last element of the database remains the same.

Copy link
Collaborator

@sbillinge sbillinge left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please see my comment.

For the bad entry, test it in a different function called test_insert_one_bad() and use a context manager that checks that the right exception is thrown. I think there are examples in the regolith tests.

try:
coll = self.dbs[dbname][collname]
coll[doc["_id"]] = doc
except TypeError:
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maybe don't catch the exception, let the program crash? Otherwise we have bad magic that the program is not updating the database but we don't know why. Here having it crash with the exception is a feature, not a bug.

except TypeError:
print('Input type should be json (dict).')
client.load_database(pydr_rc['databases'][0])
client.insert_one(dbname, collname, input)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please check we have the right logic here. I think it should look like:

  1. take exemplar and create a filesystem database with exemplar somewhere in tempdir
  2. connect to this db
  3. insert_one into this db
  4. read the file from the tempdir file
  5. make sure it contains the original exemplar and additionally the thing that was added.

I don't see the read so I am not sure we are testing that the file was updated.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think since we are calling the rc fixture when running the test, the make_db and rc functions in conftest.py automatically generates a filesystem database with exemplar in tempdir. So if we print(client.dbs) after client.insert_one, the database contains {'local': {'calculated': {...EXEMPLAR, INSERTED_DATA}, {'measured': {...EXEMPLAR}}}}. Also this is what I originally intended to do when I wrote assert list(client.dbs[dbname][collname].values())[-1] == result.

Copy link
Collaborator

@sbillinge sbillinge left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please see comments.


client.load_database(pydr_rc['databases'][0])

with pytest.raises(KeyError):
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

do we want to return more info back to the user, for example, printing what the bad entry was? To help with debugging? I don't remember the syntax but there is a way of capturing the error message and doing an assert on it.

@@ -134,6 +133,22 @@ def test_insert_one(make_db, rc, input, result):
assert list(client.dbs[dbname][collname].values())[-1] == result
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

per your question (which I don't see here for some reason) this may be left over from Regolith where the process is to load the entire database into memory (as client.dbs[dbname][collname]). It is then put back into the database at the end with a db.dump() or sthg like that. But I think we don't want to do that here as we will have 100,000 entries in the db. so we want insert_one to insert it all the way into the database backend. In the filesystem case this will be a text file on disc containing yml or json. in the mongo case it will be a remote mongodb and for mpcontribs it will be in mpcontribs.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sorry, quick questions here. So it's a bad practice to load client.dbs[dbname][collname] entirely using the above.> What if we implement find_one first and use it to test insert_one? For example: assert client.find_one(inserted_cif_id) == inserted_cif?

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we could, but I am not convinced that it is the best way. We should presumably also test that nothing else in the file was corrupted or overwritten or sthg.

Also, as I mentioned before, we are not inserting cifs per se. I think insert_one should be able to insert any valid json. I think you agree, but by putting cif in the name it makes the test harder to read so I suggest changing cif to json in the name.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you! Also another quick question, by inserting all the way into the database backend you mean overriding the yaml file directly instead of using client.load_db and client.dump_db? That means we have to change our insert_one function to directly access the yaml file?

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yes, I think that is right. Regolith assumed small databases that could be stored in memory, then operations on the items could be done with do loops and list comprehensions, but with large databases we want to use the powerful database capabilities directly, so we want kind of CRUD capabilities. We may probably want the fs to be json format not yml, but otherwise, yes, I think we want to insert directly.

with pytest.raises(TypeError):
client.insert_one(dbname, collname, test_insert_cif_bad[1])
assert '_id' in str(excinfo.value)
print(excinfo)
Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Also, how do we return info back to the user in pytest? Is print() enough? I couldn't find it on pytest docs...

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we don't want to return anything back to the user in the tests, we just want them to pass or fail. This pattern could work

def test_insert_one_bad():
    with pytest.raises(TypeError, match=r"Bad value in database entry key bad_entry_key"):
        insert_one(input[0])

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

then we need logic in the function itself so it raises an exception with the right error message.

@sl5035

This comment was marked as off-topic.

@sbillinge
Copy link
Collaborator

sbillinge commented Mar 27, 2023 via email

@sl5035
Copy link
Author

sl5035 commented Mar 28, 2023

Sorry for the delay, but still thinking about the logic here. So the goal is to modify our database (preferably in .json format) without loading the entire database into our memory, right? I thought one way to achieve that is to open the file using json.load() and append our doc in the file and dump it without storing it in self.dbs. I am still figuring out the code layout though.

@sbillinge
Copy link
Collaborator

Sorry for the delay, but still thinking about the logic here. So the goal is to modify our database (preferably in .json format) without loading the entire database into our memory, right? I thought one way to achieve that is to open the file using json.load() and append our doc in the file and dump it without storing it in self.dbs. I am still figuring out the code layout though.

yes, I think that can work. It of course makes the filesystem backend very slow, if we load and dump for every insert, but it should make the mongo backend fast because we only insert what we want to insert.

Unless there is a way to have our cake and eat it. If we detect fs backend we load at the beginning, insert everything into a memory object like self.dbs then dump at the end. It may not matter if the client behaves differently in this way if it is transparent to the front end. It will make a difference if the program crashes in the middle. I am not sure how DB programs handle this tbh.

What do you think? I don't have all the answers here

Copy link
Collaborator

@sbillinge sbillinge left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please see my comment. I wonder if we want to have a quick call and do a bit of a hackathon here? It is hard because I am not sure what the right structure is either, so we may need to play around a bit together?

@@ -234,7 +239,7 @@ def all_documents(self, collname, copy=True):
return deepcopy(self.chained_db.get(collname, {})).values()
return self.chained_db.get(collname, {}).values()

def insert_one(self, filename, dbname, collname, doc):
def insert_one(self, filename, doc):
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

do we know why this takes filename and doesn't take dbname and collname? Doesn't htis break the api, because the insert_one will have a different signature for the different clients (mongo vs fs).

I may be wrong here, so let me know if so, but I htink the idea is that insert_one has the same signature regardless of the client, but it has different behavior depending on the backend. In general, we want to insert something into a particular collection in a particular database. If that collection is in a mongo db the client will need a URL and login credentials and so on of the db, if it is in a file on a filesystem it will need to know the file-name and where in the directory structure it is located, but this is handled by the rc.databases at runtime.

@sl5035
Copy link
Author

sl5035 commented Mar 30, 2023 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants