func create_label_studio_project()
Table.
- A tutorial notebook with fully worked examples can be found here: Using Label Studio for Annotations with Pixeltable
label_config specifies the Label Studio project configuration, in XML format, as described in the Label Studio documentation. The linked project will have one column for each data field in the configuration; for example, if the configuration has an entry
<Image name=“image_obj” value=“$image”/>
then the linked project will have a column named image. In addition, the linked project will always have a JSON-typed column annotations representing the output.
By default, Pixeltable will link each of these columns to a column of the specified Table with the same name. If any of the data fields are missing, an exception will be raised. If the annotations column is missing, it will be created. The default names can be overridden by specifying an optional col_mapping, with Pixeltable column names as keys and Label Studio field names as values. In all cases, the Pixeltable columns must have types that are consistent with their corresponding Label Studio fields; otherwise, an exception will be raised.
The API key and URL for a valid Label Studio server must be specified in Pixeltable config. Either:
- Set the
LABEL_STUDIO_API_KEYandLABEL_STUDIO_URLenvironment variables; or - Specify
api_keyandurlfields in thelabel-studiosection of$PIXELTABLE_HOME/config.toml.
pip install label-studio-sdkpip install boto3(if using S3 import storage)
t(Table): The table to link to.label_config(str): The Label Studio project configuration, in XML format.name(str | None): An optional name for the new project in Pixeltable. If specified, must be a valid Pixeltable identifier and must not be the name of any other external data store linked tot. If not specified, a default name will be used of the formls_project_0,ls_project_1, etc.title(str | None): An optional title for the Label Studio project. This is the title that annotators will see inside Label Studio. Unlikename, it does not need to be an identifier and does not need to be unique. If not specified, the table namet.namewill be used.media_import_method(Literal['post', 'file', 'url'], default:post): The method to use when transferring media files to Label Studio:post: Media will be sent to Label Studio via HTTP post. This should generally only be used for prototyping; due to restrictions in Label Studio, it can only be used with projects that have just one data field, and does not scale well.file: Media will be sent to Label Studio as a file on the local filesystem. This method can be used if Pixeltable and Label Studio are running on the same host.url: Media will be sent to Label Studio as externally accessible URLs. This method cannot be used with local media files or with media generated by computed columns. The default ispost.
col_mapping(dict[str, str] | None): An optional mapping of local column names to Label Studio fields.sync_immediately(bool, default:True): IfTrue, immediately perform an initial synchronization by exporting all rows of the table as Label Studio tasks.s3_configuration(dict[str, Any] | None): If specified, S3 import storage will be configured for the new project. This can only be used withmedia_import_method='url', and ifmedia_import_method='url'and any of the media data is referenced bys3://URLs, then it must be specified in order for such media to display correctly in the Label Studio interface. The items in thes3_configurationdictionary correspond to kwarg parameters of the Label Studioconnect_s3_import_storagemethod, as described in the Label Studio connect_s3_import_storage docs.bucketmust be specified; all other parameters are optional. If credentials are not specified explicitly, Pixeltable will attempt to retrieve them from the environment (such as from~/.aws/credentials). If a title is not specified, Pixeltable will use the default'Pixeltable-S3-Import-Storage'. All other parameters use their Label Studio defaults.kwargs(Any): Additional keyword arguments are passed to thestart_projectmethod in the Label Studio SDK, as described in the Label Studio start_project docs.
UpdateStatus: AnUpdateStatusrepresenting the status of any synchronization operations that occurred.
video_col column of the table tbl:
media_import_method='url', whose media are stored in an S3 bucket:
func export_images_as_fo_dataset()
FrameIterator) will first be written to disk in the specified image_format.
The label parameters accept one or more sets of labels of each type. If a single Expr is provided, then it will be exported as a single set of labels with a default name such as classifications. (The single set of labels may still containing multiple individual labels; see below.) If a list of Exprs is provided, then each one will be exported as a separate set of labels with a default name such as classifications, classifications_1, etc. If a dictionary of Exprs is provided, then each entry will be exported as a set of labels with the specified name.
Requirements:
pip install fiftyone
tbl(pxt.Table): The table from which to export data.images(exprs.Expr): A column or expression that contains the images to export.image_format(str, default:webp): The format to use when writing out images for export.classifications(exprs.Expr | list[exprs.Expr] | dict[str, exprs.Expr] | None): Optional image classification labels. If a singleExpris provided, it must be a table column or an expression that evaluates to a list of dictionaries. Each dictionary in the list corresponds to an image class and must have the following structure:If multipleExprs are provided, each one must evaluate to a list of such dictionaries.detections(exprs.Expr | list[exprs.Expr] | dict[str, exprs.Expr] | None): Optional image detection labels. If a singleExpris provided, it must be a table column or an expression that evaluates to a list of dictionaries. Each dictionary in the list corresponds to an image detection, and must have the following structure:If multipleExprs are provided, each one must evaluate to a list of such dictionaries.
'fo.Dataset': A Voxel51 dataset.
image column of the table tbl as a Voxel51 dataset, using classification labels from tbl.classifications:
func export_lancedb()
RecordBatches, the size of which can be controlled with the batch_size_bytes parameter.
Requirements:
pip install lancedb
table_or_df(Any): Table or Dataframe to export.db_uri(Path): Local Path to the LanceDB database.table_name(Any): Name of the table in the LanceDB database.batch_size_bytes(Any): Maximum size in bytes for each batch.if_exists(Literal['error', 'overwrite', 'append'], default:error): Determines the behavior if the table already exists. Must be one of the following:'error': raise an error'overwrite': overwrite the existing table'append': append to the existing table
func export_parquet()
table_or_df(Any): Table or Dataframe to export.parquet_path(Any): Path to directory to write the parquet files to.partition_size_bytes(Any): The maximum target size for each chunk. Default 100_000_000 bytes.inline_images(Any): If True, images are stored inline in the parquet file. This is useful for small images, to be imported as pytorch dataset. But can be inefficient for large images, and cannot be imported into pixeltable. If False, will raise an error if the Dataframe has any image column. Default False.
func import_csv()
import_pandas(table_path, pd.read_csv(filepath_or_buffer, **kwargs), schema=schema). See the Pandas documentation for read_csv for more details.
Returns:
pixeltable.catalog.table.Table: A handle to the newly createdTable.
func import_excel()
import_pandas(table_path, pd.read_excel(io, *args, **kwargs), schema=schema). See the Pandas documentation for read_excel for more details.
Returns:
pixeltable.catalog.table.Table: A handle to the newly createdTable.
func import_huggingface_dataset()
datasets library to be installed.
Parameters:
table_path(str): Path to the table.dataset(datasets.Dataset | datasets.DatasetDict): Huggingfacedatasets.Datasetordatasets.DatasetDictto insert into the table.schema_overrides(dict[str, Any] | None): If specified, then for each (name, type) pair inschema_overrides, the column with namenamewill be given typetype, instead of being inferred from theDatasetorDatasetDict. The keys inschema_overridesshould be the column names of theDatasetorDatasetDict(whether or not they are valid Pixeltable identifiers).primary_key(str | list[str] | None): The primary key of the table (seecreate_table()).kwargs(Any): Additional arguments to pass tocreate_table. An argument ofcolumn_name_for_splitmust be provided if the source is a DatasetDict. This column name will contain the split information. If None, no split information will be stored.
pxt.Table: A handle to the newly createdTable.
func import_json()
import_data(table_path, json.loads(file_contents, **kwargs), ...), where file_contents is the contents of the specified filepath_or_url.
Parameters:
tbl_path(str): The name of the table to create.filepath_or_url(str): The path or URL of the JSON file.schema_overrides(dict[str, Any] | None): If specified, then columns inschema_overrideswill be given the specified types (seeimport_rows()).primary_key(str | list[str] | None): The primary key of the table (seecreate_table()).num_retained_versions(int, default:10): The number of retained versions of the table (seecreate_table()).comment(str, default: “): A comment to attach to the table (seecreate_table()).kwargs(Any): Additional keyword arguments to pass tojson.loads.
pxt.Table: A handle to the newly createdTable.
func import_pandas()
DataFrame, with the specified name. The schema of the table will be inferred from the DataFrame.
The column names of the new table will be identical to those in the DataFrame, as long as they are valid Pixeltable identifiers. If a column name is not a valid Pixeltable identifier, it will be normalized according to the following procedure:
- first replace any non-alphanumeric characters with underscores;
- then, preface the result with the letter ‘c’ if it begins with a number or an underscore;
- then, if there are any duplicate column names, suffix the duplicates with ‘_2’, ‘_3’, etc., in column order.
tbl_name(str): The name of the table to create.df(pandas.core.frame.DataFrame): The PandasDataFrame.schema_overrides(dict[str, typing.Any] | None): If specified, then for each (name, type) pair inschema_overrides, the column with namenamewill be given typetype, instead of being inferred from theDataFrame. The keys inschema_overridesshould be the column names of theDataFrame(whether or not they are valid Pixeltable identifiers).
pixeltable.catalog.table.Table: A handle to the newly createdTable.
func import_parquet()
table(str): Fully qualified name of the table to import the data into.parquet_path(str): Path to an individual Parquet file or directory of Parquet files.schema_overrides(dict[str, Any] | None): If specified, then for each (name, type) pair inschema_overrides, the column with namenamewill be given typetype, instead of being inferred from the Parquet dataset. The keys inschema_overridesshould be the column names of the Parquet dataset (whether or not they are valid Pixeltable identifiers).primary_key(str | list[str] | None): The primary key of the table (seecreate_table()).kwargs(Any): Additional arguments to pass tocreate_table.
pxt.Table: A handle to the newly created table.
func import_rows()
{column_name: value, ...}. Pixeltable will attempt to infer the schema of the table from the supplied data, using the most specific type that can represent all the values in a column.
If schema_overrides is specified, then for each entry (column_name, type) in schema_overrides, Pixeltable will force the specified column to the specified type (and will not attempt any type inference for that column).
All column types of the new table will be nullable unless explicitly specified as non-nullable in schema_overrides.
Parameters:
tbl_path(str): The qualified name of the table to create.rows(list[dict[str, Any]]): The list of dictionaries to import.schema_overrides(dict[str, Any] | None): If specified, then columns inschema_overrideswill be given the specified types as described above.primary_key(str | list[str] | None): The primary key of the table (seecreate_table()).num_retained_versions(int, default:10): The number of retained versions of the table (seecreate_table()).comment(str, default: “): A comment to attach to the table (seecreate_table()).
pxt.Table: A handle to the newly createdTable.