{"payload":{"header_redesign_enabled":false,"results":[{"id":"138026545","archived":false,"color":"#DA5B0B","followers":8,"has_funding_file":false,"hl_name":"falaktheoptimist/gradient_descent_optimizers","hl_trunc_description":"Hands on implementation of gradient descent based optimizers in raw python","language":"Jupyter Notebook","mirror":false,"owned_by_organization":false,"public":true,"repo":{"repository":{"id":138026545,"name":"gradient_descent_optimizers","owner_id":7119249,"owner_login":"falaktheoptimist","updated_at":"2018-12-20T08:52:52.523Z","has_issues":true}},"sponsorable":false,"topics":["python","optimizer","python3","adadelta","sgd","gradient-descent","adagrad","rmsprop","adam-optimizer","nesterov-accelerated-sgd"],"type":"Public","help_wanted_issues_count":0,"good_first_issue_issues_count":0,"starred_by_current_user":false}],"type":"repositories","page":1,"page_count":1,"elapsed_millis":70,"errors":[],"result_count":1,"facets":[],"protected_org_logins":[],"topics":null,"query_id":"","logged_in":false,"sign_up_path":"/signup?source=code_search_results","sign_in_path":"/login?return_to=https%3A%2F%2Fgithub.com%2Fsearch%3Fq%3Drepo%253Afalaktheoptimist%252Fgradient_descent_optimizers%2B%2Blanguage%253A%2522Jupyter%2BNotebook%2522","metadata":null,"csrf_tokens":{"/falaktheoptimist/gradient_descent_optimizers/star":{"post":"IBV5N1N9gvNB0YLBulrjJZkQg1mkTmhx1ZReRALe5N35VqONcjx5jCwAqjfVUEEfFxv3U71VrdffWQVphy-_UQ"},"/falaktheoptimist/gradient_descent_optimizers/unstar":{"post":"chxKnTJB7X-A3iuAAEy-V7E9aRp_0Bs0lv9n_4Ze3fmxQG0HqrF4d4gj1JDagPONGjyzKagF3c2TgtpNDpoHSg"},"/sponsors/batch_deferred_sponsor_buttons":{"post":"IHMCT1Itvo7IzuvQtXF1ed5jhiZXYFnCsyZu0oQ2If21uDSLVjjbHFnlOqbLHcEa6BikaCzNRxLzRE2bfFSjKg"}}},"title":"Repository search results"}