Skip to content

unitorch.cli.models.roberta¤

RobertaProcessor¤

Tip

core/process/roberta is the section for configuration of RobertaProcessor.

Bases: RobertaProcessor

Processor for the Roberta model.

Initialize the RobertaProcessor.

Parameters:

Name Type Description Default
vocab_path str

The path to the vocabulary file.

required
merge_path str

The path to the merge file.

required
max_seq_length int

The maximum sequence length. Defaults to 128.

128
source_type_id int

The source type ID. Defaults to 0.

0
target_type_id int

The target type ID. Defaults to 0.

0
Source code in src/unitorch/cli/models/roberta/processing.py
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
def __init__(
    self,
    vocab_path,
    merge_path,
    max_seq_length: Optional[int] = 128,
    source_type_id: Optional[int] = 0,
    target_type_id: Optional[int] = 0,
):
    """
    Initialize the RobertaProcessor.

    Args:
        vocab_path (str): The path to the vocabulary file.
        merge_path (str): The path to the merge file.
        max_seq_length (int, optional): The maximum sequence length. Defaults to 128.
        source_type_id (int, optional): The source type ID. Defaults to 0.
        target_type_id (int, optional): The target type ID. Defaults to 0.
    """
    super().__init__(
        vocab_path=vocab_path,
        merge_path=merge_path,
        max_seq_length=max_seq_length,
        source_type_id=source_type_id,
        target_type_id=target_type_id,
    )

from_core_configure classmethod ¤

from_core_configure(config, **kwargs)

Create an instance of RobertaProcessor from a core configuration.

Parameters:

Name Type Description Default
config

The core configuration.

required
**kwargs

Additional keyword arguments.

{}

Returns:

Name Type Description
RobertaProcessor

The initialized RobertaProcessor instance.

Source code in src/unitorch/cli/models/roberta/processing.py
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
@classmethod
@add_default_section_for_init("core/process/roberta")
def from_core_configure(cls, config, **kwargs):
    """
    Create an instance of RobertaProcessor from a core configuration.

    Args:
        config: The core configuration.
        **kwargs: Additional keyword arguments.

    Returns:
        RobertaProcessor: The initialized RobertaProcessor instance.
    """
    config.set_default_section("core/process/roberta")
    pretrained_name = config.getoption("pretrained_name", "roberta-base")
    vocab_path = config.getoption("vocab_path", None)
    vocab_path = pop_value(
        vocab_path,
        nested_dict_value(pretrained_roberta_infos, pretrained_name, "vocab"),
    )
    vocab_path = cached_path(vocab_path)

    merge_path = config.getoption("merge_path", None)
    merge_path = pop_value(
        merge_path,
        nested_dict_value(pretrained_roberta_infos, pretrained_name, "merge"),
    )
    merge_path = cached_path(merge_path)

    return {
        "vocab_path": vocab_path,
        "merge_path": merge_path,
    }

RobertaForClassification¤

Tip

core/model/classification/roberta is the section for configuration of RobertaForClassification.

Bases: RobertaForClassification

Roberta model for classification tasks.

Initialize the RobertaForClassification model.

Parameters:

Name Type Description Default
config_path str

The path to the model configuration file.

required
num_classes int

The number of output classes. Defaults to 1.

1
gradient_checkpointing bool

Whether to use gradient checkpointing. Defaults to False.

False
Source code in src/unitorch/cli/models/roberta/modeling.py
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
def __init__(
    self,
    config_path: str,
    num_classes: Optional[int] = 1,
    gradient_checkpointing: Optional[bool] = False,
):
    """
    Initialize the RobertaForClassification model.

    Args:
        config_path (str): The path to the model configuration file.
        num_classes (int, optional): The number of output classes. Defaults to 1.
        gradient_checkpointing (bool, optional): Whether to use gradient checkpointing. Defaults to False.
    """
    super().__init__(
        config_path=config_path,
        num_classes=num_classes,
        gradient_checkpointing=gradient_checkpointing,
    )

forward ¤

forward(
    input_ids: Tensor,
    attention_mask: Optional[Tensor] = None,
    token_type_ids: Optional[Tensor] = None,
    position_ids: Optional[Tensor] = None,
)

Perform a forward pass through the model.

Parameters:

Name Type Description Default
input_ids Tensor

Input token IDs.

required
attention_mask Tensor

Attention mask. Defaults to None.

None
token_type_ids Tensor

Token type IDs. Defaults to None.

None
position_ids Tensor

Position IDs. Defaults to None.

None

Returns:

Name Type Description
ClassificationOutputs

The classification outputs.

Source code in src/unitorch/cli/models/roberta/modeling.py
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
@autocast(device_type=("cuda" if torch.cuda.is_available() else "cpu"))
def forward(
    self,
    input_ids: torch.Tensor,
    attention_mask: Optional[torch.Tensor] = None,
    token_type_ids: Optional[torch.Tensor] = None,
    position_ids: Optional[torch.Tensor] = None,
):
    """
    Perform a forward pass through the model.

    Args:
        input_ids (torch.Tensor): Input token IDs.
        attention_mask (torch.Tensor, optional): Attention mask. Defaults to None.
        token_type_ids (torch.Tensor, optional): Token type IDs. Defaults to None.
        position_ids (torch.Tensor, optional): Position IDs. Defaults to None.

    Returns:
        ClassificationOutputs: The classification outputs.
    """
    outputs = super().forward(
        input_ids=input_ids,
        attention_mask=attention_mask,
        token_type_ids=token_type_ids,
        position_ids=position_ids,
    )
    return ClassificationOutputs(outputs=outputs)

from_core_configure classmethod ¤

from_core_configure(config, **kwargs)

Create an instance of RobertaForClassification from a core configuration.

Parameters:

Name Type Description Default
config

The core configuration.

required
**kwargs

Additional keyword arguments.

{}

Returns:

Name Type Description
RobertaForClassification

The initialized RobertaForClassification instance.

Source code in src/unitorch/cli/models/roberta/modeling.py
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
@classmethod
@add_default_section_for_init("core/model/classification/roberta")
def from_core_configure(cls, config, **kwargs):
    """
    Create an instance of RobertaForClassification from a core configuration.

    Args:
        config: The core configuration.
        **kwargs: Additional keyword arguments.

    Returns:
        RobertaForClassification: The initialized RobertaForClassification instance.
    """
    config.set_default_section("core/model/classification/roberta")
    pretrained_name = config.getoption("pretrained_name", "roberta-base")
    config_path = config.getoption("config_path", None)
    config_path = pop_value(
        config_path,
        nested_dict_value(pretrained_roberta_infos, pretrained_name, "config"),
    )

    config_path = cached_path(config_path)
    gradient_checkpointing = config.getoption("gradient_checkpointing", False)
    num_classes = config.getoption("num_classes", 1)

    inst = cls(config_path, num_classes, gradient_checkpointing)
    pretrained_weight_path = config.getoption("pretrained_weight_path", None)
    weight_path = pop_value(
        pretrained_weight_path,
        nested_dict_value(pretrained_roberta_infos, pretrained_name, "weight"),
        check_none=False,
    )
    if weight_path is not None:
        inst.from_pretrained(weight_path)

    return inst