v0.16.0 #71
OnlyDeniko
announced in
Announcements
v0.16.0
#71
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
polars
package. This is available in the following modules: data (Dataset, SequenceTokenizer, SequentialDataset) for working with transformers, metrics, preprocessing and splitters. The new format allows to achieve multiple acceleration of calculations relative to thePandas
andPySpark
dataframes. You can see more details about usage in the examples.seaborn
andmatplotlib
. Removed functionsreplay.utils.distributions.plot_item_dist
andreplay.utils.distributions.plot_user_dist
.get_all_embeddings
,set_item_embeddings_by_size
,set_item_embeddings_by_tensor
,append_item_embeddings
. You can see more details about their use in the examples.QueryEmbeddingsPredictionCallback
to get query embeddings at the inference stage in transformers. You can see more details about usage in the examples.SequenceTokenizer
andTorchSequentialDataset
. It becomes possible to use numerical features inside transformers.cardinality
inTensorSchema
. Now it is not necessary to pass thecardinality
parameter, the value will be calculated automatically.core_count
parameter toreplay.utils.session_handler.get_spark_session
. If nothing is specified, the env variablesREPLAY_SPARK_CORE_COUNT
andREPLAY_SPARK_MEMORY
are taken into account. If they are not specified, the value is set to-1
.item_count
parameter inValidationMetricsCallback
. If you are not going to calculate theCoverage
metric, then you do not need to pass this parameter.Coverage
metric onPandas
andPySpark
has been aligned.PySpark
toPandas
in some models. Added theallow_collect_to_master
parameter,False
by default.LabelEncoder
. The problem occurred when using multiple tuples with null values.This discussion was created from the release v0.16.0.
Beta Was this translation helpful? Give feedback.
All reactions