Using Data Adapters¶

This is an index of all the available data adapters, both savers and loaders. Note that some savers and loaders are the same (certain classes can handle both), but some are different. You will want to reference this when calling out to any of the following:

  1. Using save_to [or for just exposing metadata datasaver].

  2. Using load_from [or for just exposing metadata dataloader].

  3. Using materializers.

To read these tables, you want to first look at the key to determine which format you want – these should be human-readable and familiar to you. Then you’ll want to look at the types field to figure out which is the best for your case (the object you want to load from or save to).

Finally, look up the adapter params to see what parameters you can pass to the data adapters. The optional params come with their default value specified.

If you want more information, click on the module, it will send you to the code that implements it to see how the parameters are used.

As an example, say we wanted to save a pandas dataframe to a CSV file. We would first find the key csv, which would inform us that we want to call save_to.csv (or to.csv in the case of materialize). Then, we would look at the types field, finding that there is a pandas dataframe adapter. Finally, we would look at the params field, finding that we can pass path, and (optionally) sep (which we’d realize defaults to , when looking at the code).

All together, we’d end up with:

import pandas as pd
from hamilton.function_modifiers import value, save_to

@save_to.csv(path=value("my_file.csv"))
def my_data(...) -> pd.DataFrame:
    ...

For a less “abstracted” approach, where you just expose metadata from saving and loading, you can annotated your saving/loading functions to do so, e.g. analogous to the above you could do:

import pandas as pd
from hamilton.function_modifiers import datasaver

 def my_data(...) -> pd.DataFrame:
    # your function
    ...
    return _df  # return some df

@datasaver
def my_data_saver(my_data: pd.DataFrame, path: str) -> dict:
    # code to save my_data
    return {"path": path, "type": "csv", ...} # add other metadata

See dataloader for more information on how to load data and expose metadata via this more lighter weight way.

If you want to extend the @save_to or @load_from decorators, see Using Data Adapters for documentation, and the example in the repository for an example of how to do so.

Note that you will need to call registry.register_adapters (or import a module that does that) prior to dynamically referring to these in the code – otherwise we won’t know about them, and won’t be able to access that key!

Data Loaders¶

key

loader params

types

module

json
pathstr
dict
list
hamilton.io.default_data_loaders
json
filepath_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
chunksizeint | None=None
compressionstr | dict[str, typing.Any] | None=infer
convert_axesbool | None=None
convert_datesbool | list[str]=True
date_unitstr | None=None
dtypeUnion=None
dtype_backendstr | None=None
encodingstr | None=None
encoding_errorsstr | None=strict
enginestr=ujson
keep_default_datesbool=True
linesbool=False
nrowsint | None=None
orientstr | None=None
precise_floatbool=False
storage_optionsdict[str, typing.Any] | None=None
typstr=frame
DataFrame
hamilton.plugins.pandas_extensions
json
sourcestr | pathlib.Path | io.IOBase | bytes
schemacollections.abc.Mapping[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]] | collections.abc.Sequence[str | tuple[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]]]=None
schema_overridescollections.abc.Mapping[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]] | collections.abc.Sequence[str | tuple[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]]]=None
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
json
pathstr | bytearray | os.PathLike
XGBModel
Booster
hamilton.plugins.xgboost_extensions
literal
valueAny
Any
hamilton.io.default_data_loaders
file
pathstr
encodingstr=utf-8
str
hamilton.io.default_data_loaders
file
pathstr | pathlib.Path
LGBMModel
Booster
CVBooster
hamilton.plugins.lightgbm_extensions
pickle
pathstr
object
Any
hamilton.io.default_data_loaders
pickle
filepath_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader=None
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader=None
compressionstr | dict[str, typing.Any] | None=infer
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
environment
namestuple
dict
hamilton.io.default_data_loaders
yaml
pathstr | pathlib.Path
str
int
float
bool
dict
list
hamilton.plugins.yaml_extensions
npy
pathstr | pathlib.Path | typing.IO
mmap_modestr | None=None
allow_picklebool | None=None
fix_importsbool | None=None
encodingLiteral=ASCII
ndarray
hamilton.plugins.numpy_extensions
csv
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
sepstr | None=,
delimiterstr | None=None
headerUnion=infer
namescollections.abc.Sequence | None=None
index_colUnion=None
usecolslist[collections.abc.Hashable] | collections.abc.Callable | tuple | None=None
dtypeUnion=None
engineOptional=None
converterscollections.abc.Mapping | None=None
true_valueslist | None=None
false_valueslist | None=None
skipinitialspacebool | None=False
skiprowslist[int] | int | collections.abc.Callable[[collections.abc.Hashable], bool] | None=None
skipfooterint=0
nrowsint | None=None
na_valuescollections.abc.Hashable | collections.abc.Iterable | collections.abc.Mapping | None=None
keep_default_nabool=True
na_filterbool=True
verbosebool=False
skip_blank_linesbool=True
parse_datesbool | collections.abc.Sequence | None=False
keep_date_colbool=False
date_formatstr | None=None
dayfirstbool=False
cache_datesbool=True
iteratorbool=False
chunksizeint | None=None
compressionUnion=infer
thousandsstr | None=None
decimalstr=.
lineterminatorstr | None=None
quotecharstr | None=None
quotingint=0
doublequotebool=True
escapecharstr | None=None
commentstr | None=None
encodingstr=utf-8
encoding_errorsUnion=strict
dialectstr | csv.Dialect | None=None
on_bad_linesUnion=error
delim_whitespacebool=False
low_memorybool=True
memory_mapbool=False
float_precisionOptional=None
storage_optionsdict[str, typing.Any] | None=None
dtype_backendLiteral=numpy_nullable
DataFrame
hamilton.plugins.pandas_extensions
csv
filestr | typing.TextIO | _io.BytesIO | pathlib.Path | typing.BinaryIO | bytes
has_headerbool=True
include_headerbool=True
columnscollections.abc.Sequence[int] | collections.abc.Sequence[str]=None
new_columnsSequence=None
separatorstr=,
comment_charstr=None
quote_charstr="
skip_rowsint=0
dtypescollections.abc.Mapping[str, typing.Any] | collections.abc.Sequence[typing.Any]=None
null_valuesstr | collections.abc.Sequence[str] | dict[str, str]=None
missing_utf8_is_empty_stringbool=False
ignore_errorsbool=False
try_parse_datesbool=False
n_threadsint=None
infer_schema_lengthint=100
batch_sizeint=8192
n_rowsint=None
encodingUnion=utf8
low_memorybool=False
rechunkbool=True
use_pyarrowbool=False
storage_optionsdict=None
skip_rows_after_headerint=0
row_count_namestr=None
row_count_offsetint=0
sample_sizeint=1024
eol_charstr=
raise_if_emptybool=True
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
csv
filestr | typing.TextIO | _io.BytesIO | pathlib.Path | typing.BinaryIO | bytes
has_headerbool=True
columnscollections.abc.Sequence[int] | collections.abc.Sequence[str]=None
new_columnsSequence=None
separatorstr=,
comment_charstr=None
quote_charstr="
skip_rowsint=0
dtypescollections.abc.Mapping[str, typing.Any] | collections.abc.Sequence[typing.Any]=None
null_valuesstr | collections.abc.Sequence[str] | dict[str, str]=None
missing_utf8_is_empty_stringbool=False
ignore_errorsbool=False
try_parse_datesbool=False
n_threadsint=None
infer_schema_lengthint=100
batch_sizeint=8192
n_rowsint=None
encodingtype | str=utf8
low_memorybool=False
rechunkbool=True
use_pyarrowbool=False
storage_optionsdict=None
skip_rows_after_headerint=0
row_count_namestr=None
row_count_offsetint=0
eol_charstr=
raise_if_emptybool=True
LazyFrame
hamilton.plugins.polars_lazyframe_extensions
csv
sparkSparkSession
pathstr
headerbool=True
sepstr=,
DataFrame
hamilton.plugins.spark_extensions
parquet
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
engineLiteral=auto
columnslist[str] | None=None
storage_optionsdict[str, typing.Any] | None=None
use_nullable_dtypesbool=False
dtype_backendLiteral=numpy_nullable
filesystemstr | None=None
filterslist[tuple] | list[list[tuple]] | None=None
DataFrame
hamilton.plugins.pandas_extensions
parquet
filestr | typing.TextIO | _io.BytesIO | pathlib.Path | typing.BinaryIO | bytes
columnslist[int] | list[str]=None
n_rowsint=None
use_pyarrowbool=False
memory_mapbool=True
storage_optionsdict=None
parallelAny=auto
row_count_namestr=None
row_count_offsetint=0
low_memorybool=False
pyarrow_optionsdict=None
use_statisticsbool=True
rechunkbool=True
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
parquet
filestr | typing.TextIO | _io.BytesIO | pathlib.Path | typing.BinaryIO | bytes
columnslist[int] | list[str]=None
n_rowsint=None
use_pyarrowbool=False
memory_mapbool=True
storage_optionsdict=None
parallelAny=auto
row_count_namestr=None
row_count_offsetint=0
low_memorybool=False
use_statisticsbool=True
rechunkbool=True
LazyFrame
hamilton.plugins.polars_lazyframe_extensions
parquet
sparkSparkSession
pathstr
DataFrame
hamilton.plugins.spark_extensions
sql
query_or_tablestr
db_connectionstr | sqlite3.Connection
chunksizeint | None=None
coerce_floatbool=True
columnslist[str] | None=None
dtypeUnion=None
dtype_backendstr | None=None
index_colstr | list[str] | None=None
paramslist | tuple | dict | None=None
parse_dateslist | dict | None=None
DataFrame
hamilton.plugins.pandas_extensions
xml
path_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
xpathstr | None=./*
namespacedict[str, str] | None=None
elems_onlybool | None=False
attrs_onlybool | None=False
nameslist[str] | None=None
dtypedict[str, typing.Any] | None=None
convertersdict[int | str, typing.Any] | None=None
parse_datesbool | list[int | str | list[list] | dict[str, list[int]]]=False
encodingstr | None=utf-8
parserstr=lxml
stylesheetstr | pathlib.Path | _io.BytesIO | _io.BufferedReader=None
iterparsedict[str, list[str]] | None=None
compressionstr | dict[str, typing.Any] | None=infer
storage_optionsdict[str, typing.Any] | None=None
dtype_backendstr=numpy_nullable
DataFrame
hamilton.plugins.pandas_extensions
html
iostr | pathlib.Path | _io.BytesIO | _io.BufferedReader
matchstr | None=.+
flavorstr | collections.abc.Sequence | None=None
headerint | collections.abc.Sequence | None=None
index_colint | collections.abc.Sequence | None=None
skiprowsint | collections.abc.Sequence | slice | None=None
attrsdict[str, str] | None=None
parse_datesbool | None=None
thousandsstr | None=,
encodingstr | None=None
decimalstr=.
convertersdict[typing.Any, typing.Any] | None=None
na_valuesIterable=None
keep_default_nabool=True
displayed_onlybool=True
extract_linksOptional=None
dtype_backendLiteral=numpy_nullable
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
stata
filepath_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
convert_datesbool=True
convert_categoricalsbool=True
index_colstr | None=None
convert_missingbool=False
preserve_dtypesbool=True
columnscollections.abc.Sequence | None=None
order_categoricalsbool=True
chunksizeint | None=None
iteratorbool=False
compressionUnion=infer
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
feather
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
columnscollections.abc.Sequence | None=None
use_threadsbool=True
storage_optionsdict[str, typing.Any] | None=None
dtype_backendLiteral=numpy_nullable
DataFrame
hamilton.plugins.pandas_extensions
feather
sourcestr | typing.BinaryIO | _io.BytesIO | pathlib.Path | bytes
columnslist[str] | list[int] | None=None
n_rowsint | None=None
use_pyarrowbool=False
memory_mapbool=True
storage_optionsdict[str, typing.Any] | None=None
row_count_namestr | None=None
row_count_offsetint=0
rechunkbool=True
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
feather
sourcestr | typing.BinaryIO | _io.BytesIO | pathlib.Path | bytes
columnslist[str] | list[int] | None=None
n_rowsint | None=None
use_pyarrowbool=False
memory_mapbool=True
storage_optionsdict[str, typing.Any] | None=None
row_count_namestr | None=None
row_count_offsetint=0
rechunkbool=True
LazyFrame
hamilton.plugins.polars_lazyframe_extensions
orc
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
columnslist[str] | None=None
dtype_backendLiteral=numpy_nullable
filesystempyarrow._fs.FileSystem | fsspec.spec.AbstractFileSystem | None=None
DataFrame
hamilton.plugins.pandas_extensions
excel
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader=None
sheet_namestr | int | list[int | str] | None=0
headerint | collections.abc.Sequence | None=0
namescollections.abc.Sequence | None=None
index_colint | str | collections.abc.Sequence | None=None
usecolsint | str | collections.abc.Sequence | collections.abc.Callable[[str], bool] | None=None
dtypeUnion=None
engineOptional=None
convertersdict[str, collections.abc.Callable] | dict[int, collections.abc.Callable] | None=None
true_valuescollections.abc.Iterable | None=None
false_valuescollections.abc.Iterable | None=None
skiprowscollections.abc.Sequence | int | collections.abc.Callable[[int], object] | None=None
nrowsint | None=None
keep_default_nabool=True
na_filterbool=True
verbosebool=False
parse_dateslist[int | str] | dict[str, list[int | str]] | bool=False
date_formatdict[collections.abc.Hashable, str] | str | None=None
thousandsstr | None=None
decimalstr=.
commentstr | None=None
skipfooterint=0
storage_optionsdict[str, typing.Any] | None=None
dtype_backendLiteral=numpy_nullable
engine_kwargsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
table
filepath_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
sepstr | None=None
delimiterstr | None=None
headerint | collections.abc.Sequence | str | None=infer
namescollections.abc.Sequence | None=None
index_colint | str | collections.abc.Sequence | None=None
usecolscollections.abc.Sequence | None=None
dtypeUnion=None
engineOptional=None
convertersdict[collections.abc.Hashable, collections.abc.Callable] | None=None
true_valuescollections.abc.Iterable | None=None
false_valuescollections.abc.Iterable | None=None
skipinitialspacebool=False
skiprowslist[int] | int | list[collections.abc.Callable] | None=None
skipfooterint=0
nrowsint | None=None
na_valuescollections.abc.Hashable | collections.abc.Iterable | dict[collections.abc.Hashable, collections.abc.Iterable] | None=None
keep_default_nabool=True
na_filterbool=True
verbosebool=False
skip_blank_linesbool=True
parse_dateslist[int | str] | dict[str, list[int | str]] | bool=False
infer_datetime_formatbool=False
keep_date_colbool=False
date_parsercollections.abc.Callable | None=None
date_formatstr | None=None
dayfirstbool=False
cache_datesbool=True
iteratorbool=False
chunksizeint | None=None
compressionstr | dict=infer
thousandsstr | None=None
decimalstr=.
lineterminatorstr | None=None
quotecharstr | None="
quotingint=0
doublequotebool=True
escapecharstr | None=None
commentstr | None=None
encodingstr | None=None
encoding_errorsstr | None=strict
dialectstr | None=None
on_bad_linesUnion=error
delim_whitespacebool=False
low_memorybool=True
memory_mapbool=False
float_precisionOptional=None
storage_optionsdict | None=None
dtype_backendLiteral=numpy_nullable
DataFrame
hamilton.plugins.pandas_extensions
fwf
filepath_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
colspecsstr | list[tuple[int, int]] | tuple[int, int]=infer
widthslist[int] | None=None
infer_nrowsint=100
dtype_backendLiteral=numpy_nullable
DataFrame
hamilton.plugins.pandas_extensions
spss
pathstr | pathlib.Path
usecolslist[collections.abc.Hashable] | collections.abc.Callable[[str], bool] | None=None
convert_categoricalsbool=True
dtype_backendLiteral=numpy_nullable
DataFrame
hamilton.plugins.pandas_extensions
avro
filestr | typing.TextIO | _io.BytesIO | pathlib.Path | typing.BinaryIO | bytes
columnslist[int] | list[str] | None=None
n_rowsint | None=None
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
ndjson
sourcestr | pathlib.Path | io.IOBase | bytes
schemacollections.abc.Mapping[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]] | collections.abc.Sequence[str | tuple[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]]]=None
schema_overridescollections.abc.Mapping[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]] | collections.abc.Sequence[str | tuple[str, typing.Union[ForwardRef('DataTypeClass'), ForwardRef('DataType'), type[int], type[float], type[bool], type[str], type['date'], type['time'], type['datetime'], type['timedelta'], type[list[typing.Any]], type[tuple[typing.Any, ...]], type[bytes], type[object], type['Decimal'], type[None], NoneType]]]=None
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
database
querystr
connectionUnion
iter_batchesbool=False
batch_sizeint | None=None
schema_overridesdict[str, typing.Any] | None=None
infer_schema_lengthint | None=None
execute_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
spreadsheet
sourcestr | pathlib.Path | io.IOBase | bytes
sheet_idint | collections.abc.Sequence[int] | None=None
sheet_namestr | list[str] | tuple[str] | None=None
engineLiteral=xlsx2csv
engine_optionsdict[str, typing.Any] | None=None
read_optionsdict[str, typing.Any] | None=None
schema_overridesdict[str, typing.Any] | None=None
raise_if_emptybool=True
DataFrame
hamilton.plugins.polars_post_1_0_0_extensions
dlt
resourceDltResource
DataFrame
hamilton.plugins.dlt_extensions
mlflow
model_uristr | None=None
modeLiteral=tracking
run_idstr | None=None
pathstr | pathlib.Path=model
model_namestr | None=None
versionstr | int | None=None
version_aliasstr | None=None
flavormodule | str | None=None
mlflow_kwargsdict=None
Any
hamilton.plugins.mlflow_extensions

Data Savers¶

key

saver params

types

module

json
pathstr
dict
list
hamilton.io.default_data_loaders
json
filepath_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
compressionstr=infer
date_formatstr=epoch
date_unitstr=ms
default_handlercollections.abc.Callable[[typing.Any], str | float | bool | list | dict | None] | None=None
double_precisionint=10
force_asciibool=True
indexbool | None=None
indentint=0
linesbool=False
modestr=w
orientstr | None=None
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
json
fileio.IOBase | str | pathlib.Path
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
json
pathstr | os.PathLike
XGBModel
Booster
hamilton.plugins.xgboost_extensions
file
pathstr
encodingstr=utf-8
str
hamilton.io.default_data_loaders
file
pathpathlib.Path | str
bytes
BytesIO
hamilton.io.default_data_loaders
file
pathstr | pathlib.Path
num_iterationint | None=None
start_iterationint=0
importance_typeLiteral=split
LGBMModel
Booster
CVBooster
hamilton.plugins.lightgbm_extensions
pickle
pathstr
object
hamilton.io.default_data_loaders
pickle
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
compressionstr | dict[str, typing.Any] | None=infer
protocolint=5
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
memory
Any
hamilton.io.default_data_loaders
yaml
pathstr | pathlib.Path
str
int
float
bool
dict
list
hamilton.plugins.yaml_extensions
plt
pathstr | os.PathLike | typing.IO
dpifloat | str | None=None
formatstr | None=None
metadatadict | None=None
bbox_inchesstr | matplotlib.transforms.Bbox | None=None
pad_inchesfloat | str | None=None
facecolorstr | float | tuple | None=None
edgecolorstr | float | tuple | None=None
backendstr | None=None
orientationstr | None=None
papertypestr | None=None
transparentbool | None=None
bbox_extra_artistslist[matplotlib.artist.Artist] | None=None
pil_kwargsdict | None=None
Figure
hamilton.plugins.matplotlib_extensions
npy
pathstr | pathlib.Path | typing.IO
allow_picklebool | None=None
fix_importsbool | None=None
ndarray
hamilton.plugins.numpy_extensions
csv
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
sepstr | None=,
na_repstr=
float_formatstr | collections.abc.Callable | None=None
columnscollections.abc.Sequence | None=None
headerbool | list[str] | None=True
indexbool | None=False
index_labelcollections.abc.Hashable | collections.abc.Iterator[collections.abc.Hashable] | None=None
modestr=w
encodingstr | None=None
compressionUnion=infer
quotingint | None=None
quotecharstr | None="
lineterminatorstr | None=None
chunksizeint | None=None
date_formatstr | None=None
doublequotebool=True
escapecharstr | None=None
decimalstr=.
errorsstr=strict
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
csv
file_io.BytesIO | _io.TextIOWrapper | str | pathlib.Path
include_headerbool=True
separatorstr=,
line_terminatorstr=
quote_charstr="
batch_sizeint=1024
datetime_formatstr=None
date_formatstr=None
time_formatstr=None
float_precisionint=None
null_valuestr=None
quote_styletype=None
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
parquet
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
engineLiteral=auto
compressionstr | None=snappy
indexbool | None=None
partition_colslist[str] | None=None
storage_optionsdict[str, typing.Any] | None=None
extra_kwargsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
parquet
file_io.BytesIO | _io.TextIOWrapper | str | pathlib.Path
compressionAny=zstd
compression_levelint=None
statisticsbool=False
row_group_sizeint=None
use_pyarrowbool=False
pyarrow_optionsdict=None
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
sql
table_namestr
db_connectionAny
chunksizeint | None=None
dtypeUnion=None
if_existsstr=fail
indexbool=True
index_labelcollections.abc.Hashable | collections.abc.Iterator[collections.abc.Hashable] | None=None
methodstr | collections.abc.Callable | None=None
schemastr | None=None
DataFrame
hamilton.plugins.pandas_extensions
xml
path_or_bufferstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
indexbool=True
root_namestr=data
row_namestr=row
na_repstr | None=None
attr_colslist[str] | None=None
elems_colslist[str] | None=None
namespacesdict[str, str] | None=None
prefixstr | None=None
encodingstr=utf-8
xml_declarationbool=True
pretty_printbool=True
parserstr=lxml
stylesheetstr | pathlib.Path | _io.BytesIO | _io.BufferedReader | None=None
compressionstr | dict[str, typing.Any] | None=infer
storage_optionsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
html
bufstr | pathlib.Path | _io.StringIO | None=None
columnslist[str] | None=None
col_spacestr | int | list | dict | None=None
headerbool | None=True
indexbool | None=True
na_repstr | None=NaN
formatterslist | tuple | dict | None=None
float_formatstr | None=None
sparsifybool | None=True
index_namesbool | None=True
justifystr=None
max_rowsint | None=None
max_colsint | None=None
show_dimensionsbool=False
decimalstr=.
bold_rowsbool=True
classesstr | list[str] | tuple | None=None
escapebool | None=True
notebookLiteral=False
borderint=None
table_idstr | None=None
render_linksbool=False
encodingstr | None=utf-8
DataFrame
hamilton.plugins.pandas_extensions
stata
pathstr | pathlib.Path | _io.BufferedReader=None
convert_datesdict[collections.abc.Hashable, str] | None=None
write_indexbool=True
byteorderstr | None=None
time_stampdatetime.datetime | None=None
data_labelstr | None=None
variable_labelsdict[collections.abc.Hashable, str] | None=None
versionLiteral=114
convert_strlstr | None=None
compressionUnion=infer
storage_optionsdict[str, typing.Any] | None=None
value_labelsdict[collections.abc.Hashable, str] | None=None
DataFrame
hamilton.plugins.pandas_extensions
feather
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
deststr | None=None
compressionLiteral=None
compression_levelint | None=None
chunksizeint | None=None
versionint | None=2
DataFrame
hamilton.plugins.pandas_extensions
feather
filetyping.BinaryIO | _io.BytesIO | str | pathlib.Path | None=None
compressiontype=uncompressed
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
orc
pathstr | pathlib.Path | _io.BytesIO | _io.BufferedReader
engineLiteral=pyarrow
indexbool | None=None
engine_kwargsdict[str, typing.Any] | None=None
DataFrame
hamilton.plugins.pandas_extensions
excel
pathstr | pathlib.Path | _io.BytesIO
sheet_namestr=Sheet1
na_repstr=
float_formatstr | None=None
columnscollections.abc.Sequence | None=None
headercollections.abc.Sequence | bool=True
indexbool=True
index_labelcollections.abc.Hashable | collections.abc.Iterator[collections.abc.Hashable] | None=None
startrowint=0
startcolint=0
engineOptional=None
merge_cellsbool=True
inf_repstr=inf
freeze_panestuple[int, int] | None=None
storage_optionsdict[str, typing.Any] | None=None
engine_kwargsdict[str, typing.Any] | None=None
modeOptional=w
if_sheet_existsOptional=None
datetime_formatstr=None
date_formatstr=None
DataFrame
hamilton.plugins.pandas_extensions
avro
file_io.BytesIO | _io.TextIOWrapper | str | pathlib.Path
compressionAny=uncompressed
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
ndjson
fileio.IOBase | str | pathlib.Path
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
database
table_namestr
connectionUnion
if_table_existsLiteral=fail
engineLiteral=sqlalchemy
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
spreadsheet
workbooktype | _io.BytesIO | pathlib.Path | str
worksheetstr | None=None
positiontuple[int, int] | str=A1
table_stylestr | dict[str, typing.Any] | None=None
table_namestr | None=None
column_formatscollections.abc.Mapping[str | tuple[str, ...], str | collections.abc.Mapping[str, str]] | None=None
dtype_formatsdict[polars.datatypes.classes.DataType | polars.datatypes.classes.DataTypeClass, str] | None=None
conditional_formatscollections.abc.Mapping[str | collections.abc.Collection[str], str | collections.abc.Mapping[str, typing.Any] | collections.abc.Sequence[str | collections.abc.Mapping[str, typing.Any]]] | None=None
header_formatdict[str, typing.Any] | None=None
column_totalscollections.abc.Mapping[typing.Union[ForwardRef('str'), ForwardRef('Selector'), tuple[typing.Union[ForwardRef('str'), ForwardRef('Selector')]]], str] | collections.abc.Sequence[str] | bool | None=None
column_widthscollections.abc.Mapping[str, tuple[str, ...] | int] | int | None=None
row_totalscollections.abc.Mapping[str, str | collections.abc.Collection[str]] | collections.abc.Collection[str] | bool | None=None
row_heightsdict[int | tuple[int, ...], int] | int | None=None
sparklinesdict[str, collections.abc.Sequence[str] | dict[str, typing.Any]] | None=None
formulasdict[str, str | dict[str, str]] | None=None
float_precisionint=3
include_headerbool=True
autofilterbool=True
autofitbool=False
hidden_columnscollections.abc.Sequence[str] | str | None=None
hide_gridlinesbool=None
sheet_zoomint | None=None
freeze_panesstr | tuple[int, int] | tuple[str, int, int] | tuple[int, int, int, int] | None=None
DataFrame
LazyFrame
hamilton.plugins.polars_post_1_0_0_extensions
png
pathstr | os.PathLike
dpifloat=200
formatstr=png
metadatadict | None=None
bbox_inchesstr=None
pad_inchesfloat=0.1
backendstr | None=None
papertypestr=None
transparentbool=None
bbox_extra_artistslist | None=None
pil_kwargsdict | None=None
ConfusionMatrixDisplay
DetCurveDisplay
PrecisionRecallDisplay
PredictionErrorDisplay
RocCurveDisplay
DecisionBoundaryDisplay
LearningCurveDisplay
PartialDependenceDisplay
ValidationCurveDisplay
Figure
hamilton.plugins.sklearn_plot_extensions
dlt
pipelinePipeline
table_namestr
primary_keystr | None=None
write_dispositionOptional=None
columnscollections.abc.Sequence[dlt.common.schema.typing.TColumnSchema] | None=None
schemadlt.common.schema.schema.Schema | None=None
loader_file_formatOptional=None
Iterable
DataFrame
Table
RecordBatch
hamilton.plugins.dlt_extensions
mlflow
pathstr | pathlib.Path=model
register_asstr | None=None
flavorstr | module | None=None
run_idstr | None=None
mlflow_kwargsdict=None
Any
hamilton.plugins.mlflow_extensions