{"payload":{"header_redesign_enabled":false,"results":[{"id":"466742571","archived":false,"color":"#DA5B0B","followers":31,"has_funding_file":false,"hl_name":"S-Abdelnabi/OoC-multi-modal-fc","hl_trunc_description":"Code for our CVPR'22 paper: Open-Domain, Content-based, Multi-modal Fact-checking of Out-of-Context Images via Online Resources","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":466742571,"name":"OoC-multi-modal-fc","owner_id":59022078,"owner_login":"S-Abdelnabi","updated_at":"2022-11-15T17:58:14.462Z","has_issues":true}},"sponsorable":false,"topics":[],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":93,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253AS-Abdelnabi%252FOoC-multi-modal-fc%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"warn_limited_results":false,"csrf_tokens":{"/S-Abdelnabi/OoC-multi-modal-fc/star":{"post":"prGB-85vcnpUocmcQFnQp9hRngNMRisyFqc_4zkd5sOVMsqCgKr47IpmHKMVPDg91HWlHaC8AJRvfDzoAQaFtA"},"/S-Abdelnabi/OoC-multi-modal-fc/unstar":{"post":"Z3-ZI_GGthQa168LObahxcuFwxWfPDkPUMZtByNvpr3Sk0fZmo8zl7a3mFgggQl_DgO4PykGNllf9bfyVcmBQQ"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"nHC-EsDi5MqoV_97mTaRqR9g3-tFzqlx6kBDSeejxq5ZRXJZRaJpg_L6Vq_iywMTC0b3vskUCYbD_GUVS94Lgw"}}},"title":"Repository search results"}