Skip to content

Commit

Permalink
No public description
Browse files Browse the repository at this point in the history
PiperOrigin-RevId: 610693234
  • Loading branch information
tf-text-github-robot committed Feb 27, 2024
1 parent daf770b commit 37e8e00
Show file tree
Hide file tree
Showing 12 changed files with 20 additions and 20 deletions.
4 changes: 2 additions & 2 deletions tensorflow_text/core/ops/fast_sentencepiece_ops.cc
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ REGISTER_OP("TFText>FastSentencepieceTokenize")
tensorflow::shape_inference::DimensionHandle num_splits;
TF_RETURN_IF_ERROR(c->Add(c->NumElements(c->input(1)), 1, &num_splits));
c->set_output(1, c->Vector(num_splits));
return tensorflow::OkStatus();
return absl::OkStatus();
});

REGISTER_OP("TFText>FastSentencepieceDetokenize")
Expand All @@ -66,7 +66,7 @@ REGISTER_OP("TFText>FastSentencepieceDetokenize")
shape_inference::DimensionHandle dim;
TF_RETURN_IF_ERROR(c->Subtract(c->NumElements(c->input(2)), 1, &dim));
c->set_output(0, c->Vector(dim));
return OkStatus();
return absl::OkStatus();
});

} // namespace text
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/mst_ops.cc
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ REGISTER_OP("MaxSpanningTree")

context->set_output(0, context->Vector(batch_size));
context->set_output(1, context->Matrix(batch_size, max_nodes));
return tensorflow::OkStatus();
return absl::OkStatus();
})
.Doc(R"doc(
Finds the maximum directed spanning tree of a digraph.
Expand Down
4 changes: 2 additions & 2 deletions tensorflow_text/core/ops/normalize_ops.cc
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ REGISTER_OP("NormalizeUTF8WithOffsetsMap")
.SetShapeFn([](::tensorflow::shape_inference::InferenceContext* c) {
c->set_output(0, c->input(0));
c->set_output(1, c->input(0));
return OkStatus();
return absl::OkStatus();
})
.Doc(R"doc(
Normalizes each UTF8 string in the input tensor using 'normalization_form'
Expand All @@ -81,7 +81,7 @@ REGISTER_OP("FindSourceOffsets")
TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 1, &unused));
TF_RETURN_IF_ERROR(c->WithRank(c->input(2), 1, &unused));
c->set_output(0, c->input(1));
return OkStatus();
return absl::OkStatus();
})
.Doc(R"doc(
Map the post-normalized string offsets in the input tensor to the pre-normalized
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/regex_split_ops.cc
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ Status RegexSplitOpShape(shape_inference::InferenceContext* c) {
for (int i = 0; i < c->num_outputs(); ++i) {
c->set_output(i, c->UnknownShapeOfRank(1));
}
return OkStatus();
return absl::OkStatus();
}

REGISTER_OP("RegexSplitWithOffsets")
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/rouge_l_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,7 @@ Status RougeLShapeFn(InferenceContext* c) {
c->set_output(1, c->Vector(dim));
c->set_output(2, c->Vector(dim));

return OkStatus();
return absl::OkStatus();
}

} // namespace tensorflow
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/sentence_breaking_ops.cc
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ Status SentenceFragmentShapeFn(
c->set_output(i, c->UnknownShapeOfRank(1));
}

return OkStatus();
return absl::OkStatus();
}

REGISTER_OP("SentenceFragments")
Expand Down
14 changes: 7 additions & 7 deletions tensorflow_text/core/ops/sentencepiece_ops.cc
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ REGISTER_OP("SentencepieceOp")
.SetIsStateful()
.SetShapeFn([](InferenceContext* c) {
c->set_output(0, c->Scalar());
return OkStatus();
return absl::OkStatus();
});

REGISTER_OP("SentencepieceTokenizeOp")
Expand Down Expand Up @@ -67,7 +67,7 @@ REGISTER_OP("SentencepieceTokenizeOp")
TF_RETURN_IF_ERROR(c->Add(c->NumElements(c->input(1)), 1, &num_splits));
c->set_output(1, c->Vector(num_splits));
}
return OkStatus();
return absl::OkStatus();
});

REGISTER_OP("SentencepieceTokenizeWithOffsetsOp")
Expand Down Expand Up @@ -107,7 +107,7 @@ REGISTER_OP("SentencepieceTokenizeWithOffsetsOp")
}
c->set_output(2, c->Vector(InferenceContext::kUnknownDim));
c->set_output(3, c->Vector(InferenceContext::kUnknownDim));
return OkStatus();
return absl::OkStatus();
});

REGISTER_OP("SentencepieceDetokenizeOp")
Expand All @@ -132,7 +132,7 @@ REGISTER_OP("SentencepieceDetokenizeOp")
shape_inference::DimensionHandle dim;
TF_RETURN_IF_ERROR(c->Subtract(c->NumElements(c->input(2)), 1, &dim));
c->set_output(0, c->Vector(dim));
return OkStatus();
return absl::OkStatus();
});

REGISTER_OP("SentencepieceVocabSizeOp")
Expand All @@ -142,7 +142,7 @@ REGISTER_OP("SentencepieceVocabSizeOp")
shape_inference::ShapeHandle unused;
TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused));
c->set_output(0, c->Scalar());
return OkStatus();
return absl::OkStatus();
});

REGISTER_OP("SentencepieceIdToStringOp")
Expand All @@ -154,7 +154,7 @@ REGISTER_OP("SentencepieceIdToStringOp")
TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused));
TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 1, &unused));
c->set_output(0, c->input(1));
return OkStatus();
return absl::OkStatus();
});

REGISTER_OP("SentencepieceStringToIdOp")
Expand All @@ -166,7 +166,7 @@ REGISTER_OP("SentencepieceStringToIdOp")
TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 0, &unused));
TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 1, &unused));
c->set_output(0, c->input(1));
return OkStatus();
return absl::OkStatus();
});

} // namespace text
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/split_merge_tokenize_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ Status SplitMergeTokenizeWithOffsetsShapeFn(InferenceContext* c) {
c->set_output(1, c->Vector(num_splits)); // row_splits
c->set_output(2, c->UnknownShapeOfRank(1)); // start_values
c->set_output(3, c->UnknownShapeOfRank(1)); // limit_values
return OkStatus();
return absl::OkStatus();
}

} // namespace text
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/tokenizer_from_logits_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -115,7 +115,7 @@ Status TokenizerFromLogitsShapeFn(InferenceContext* c) {
c->set_output(1, c->Vector(num_splits)); // row_splits
c->set_output(2, c->UnknownShapeOfRank(1)); // start_values
c->set_output(3, c->UnknownShapeOfRank(1)); // limit_values
return OkStatus();
return absl::OkStatus();
}

} // namespace text
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/unicode_script_tokenize_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ REGISTER_OP("UnicodeScriptTokenizeWithOffsets")
c->set_output(2, c->Vector(InferenceContext::kUnknownDim));
c->set_output(3, c->Vector(InferenceContext::kUnknownDim));
c->set_output(4, c->Vector(InferenceContext::kUnknownDim));
return OkStatus();
return absl::OkStatus();
});

} // namespace text
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/whitespace_tokenize_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ REGISTER_OP("WhitespaceTokenizeWithOffsets")
c->set_output(2, c->Vector(InferenceContext::kUnknownDim));
c->set_output(3, c->Vector(InferenceContext::kUnknownDim));
c->set_output(4, c->Vector(InferenceContext::kUnknownDim));
return OkStatus();
return absl::OkStatus();
});

} // namespace text
Expand Down
2 changes: 1 addition & 1 deletion tensorflow_text/core/ops/wordpiece_op.cc
Original file line number Diff line number Diff line change
Expand Up @@ -109,7 +109,7 @@ Status WordpieceTokenizeWithOffsetsShapeFn(InferenceContext* c) {
}
c->set_output(2, c->UnknownShapeOfRank(1)); // start_values
c->set_output(3, c->UnknownShapeOfRank(1)); // limit_values
return OkStatus();
return absl::OkStatus();
}


Expand Down

0 comments on commit 37e8e00

Please sign in to comment.