Conversation
Learning
- XPM Configxpmir.conversation.learning.DatasetConversationEntrySampler(*, dataset)[source]
Bases:
BaseSampler
Submit type:
xpmir.conversation.learning.DatasetConversationEntrySampler
Uses a conversation dataset and topic records entries
- dataset: datamaestro_text.data.conversation.base.ConversationDataset
The conversation dataset
- XPM Configxpmir.conversation.learning.reformulation.ConversationRepresentationEncoder[source]
Bases:
TextEncoderBase
[List
[Record
],RepresentationOutput
],ABC
Submit type:
xpmir.conversation.learning.reformulation.ConversationRepresentationEncoder
- XPM Configxpmir.conversation.learning.reformulation.DecontextualizedQueryConverter[source]
Bases:
Converter
[Record
,str
]Submit type:
xpmir.conversation.learning.reformulation.DecontextualizedQueryConverter
CoSPLADE
- XPM Configxpmir.conversation.models.cosplade.AsymetricMSEContextualizedRepresentationLoss(*, weight)[source]
Bases:
AlignmentLoss
[CoSPLADEOutput
,TextsRepresentationOutput
]Submit type:
xpmir.conversation.models.cosplade.AsymetricMSEContextualizedRepresentationLoss
Computes the asymetric loss for CoSPLADE
- weight: float = 1.0
Weight for this loss
- XPM Configxpmir.conversation.models.cosplade.CoSPLADE(*, history_size, queries_encoder, history_encoder)[source]
Bases:
ConversationRepresentationEncoder
Submit type:
xpmir.conversation.models.cosplade.CoSPLADE
CoSPLADE model
- history_size: int = 0
Size of history to take into account (0 for infinite)
- queries_encoder: xpmir.neural.splade.SpladeTextEncoderV2[List[List[str]]]
Encoder for the query history (the first one being the current one)
- history_encoder: xpmir.neural.splade.SpladeTextEncoderV2[Tuple[str, str]]
Encoder for (query, answer) pairs