snorkelflow.client.fm_suite.run_lf_inference
- snorkelflow.client.fm_suite.run_lf_inference(node, lf_uid, inference_splits=None, sync=False)
Run the LF on new data. Note, if there are any datapoints for which inference has been computed already the cached result will be used.
Parameters
Parameters
Returns
Returns
job_uid – The uid of the Warm Start job which can be used to monitor progress with sf.poll_job_status(job_uid).
Return type
Return type
str
Name Type Default Info node int
Node uid which contains the LF and data to run over. lf_uid int
The LF for which new predictions will be computed. inference_splits Optional[List[str]]
None
Dataset splits to run inference over. Defaults to all, [“train”, “dev”, “valid”, “test”]. sync bool
False
If True, method will block until the inference job is complete. Note the job progress can always be be monitored manually with sf.poll_job_status(job_uid). Examples
>>> sf.run_lf_inference(NODE_UID, LF_UID, ["train", "test"])
Note the job progress can be monitored with sf.poll_job_status('123')