Update ops-related pbtxt files.
PiperOrigin-RevId: 297734251
Change-Id: I99099509da3708b1736c81a0a9900234886cc9fb
diff --git a/tensorflow/core/ops/compat/ops_history_v1/QuantizedMatMulWithBiasAndDequantize.pbtxt b/tensorflow/core/ops/compat/ops_history_v1/QuantizedMatMulWithBiasAndDequantize.pbtxt
new file mode 100644
index 0000000..04ecfdb
--- /dev/null
+++ b/tensorflow/core/ops/compat/ops_history_v1/QuantizedMatMulWithBiasAndDequantize.pbtxt
@@ -0,0 +1,115 @@
+op {
+ name: "QuantizedMatMulWithBiasAndDequantize"
+ input_arg {
+ name: "a"
+ type_attr: "T1"
+ }
+ input_arg {
+ name: "b"
+ type_attr: "T2"
+ }
+ input_arg {
+ name: "bias"
+ type_attr: "Tbias"
+ }
+ input_arg {
+ name: "min_a"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "max_a"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "min_b"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "max_b"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "min_freezed_output"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "max_freezed_output"
+ type: DT_FLOAT
+ }
+ output_arg {
+ name: "out"
+ type_attr: "Toutput"
+ }
+ attr {
+ name: "T1"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_QINT8
+ type: DT_QUINT8
+ type: DT_QINT32
+ type: DT_QINT16
+ type: DT_QUINT16
+ }
+ }
+ }
+ attr {
+ name: "T2"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_QINT8
+ type: DT_QUINT8
+ type: DT_QINT32
+ type: DT_QINT16
+ type: DT_QUINT16
+ }
+ }
+ }
+ attr {
+ name: "Tbias"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_FLOAT
+ type: DT_QINT32
+ }
+ }
+ }
+ attr {
+ name: "Toutput"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_FLOAT
+ }
+ }
+ }
+ attr {
+ name: "transpose_a"
+ type: "bool"
+ default_value {
+ b: false
+ }
+ }
+ attr {
+ name: "transpose_b"
+ type: "bool"
+ default_value {
+ b: false
+ }
+ }
+ attr {
+ name: "input_quant_mode"
+ type: "string"
+ default_value {
+ s: "MIN_FIRST"
+ }
+ allowed_values {
+ list {
+ s: "MIN_FIRST"
+ s: "SCALED"
+ }
+ }
+ }
+}
diff --git a/tensorflow/core/ops/ops.pbtxt b/tensorflow/core/ops/ops.pbtxt
index 526a1bf..e82c271 100644
--- a/tensorflow/core/ops/ops.pbtxt
+++ b/tensorflow/core/ops/ops.pbtxt
@@ -31269,6 +31269,121 @@
}
}
op {
+ name: "QuantizedMatMulWithBiasAndDequantize"
+ input_arg {
+ name: "a"
+ type_attr: "T1"
+ }
+ input_arg {
+ name: "b"
+ type_attr: "T2"
+ }
+ input_arg {
+ name: "bias"
+ type_attr: "Tbias"
+ }
+ input_arg {
+ name: "min_a"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "max_a"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "min_b"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "max_b"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "min_freezed_output"
+ type: DT_FLOAT
+ }
+ input_arg {
+ name: "max_freezed_output"
+ type: DT_FLOAT
+ }
+ output_arg {
+ name: "out"
+ type_attr: "Toutput"
+ }
+ attr {
+ name: "T1"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_QINT8
+ type: DT_QUINT8
+ type: DT_QINT32
+ type: DT_QINT16
+ type: DT_QUINT16
+ }
+ }
+ }
+ attr {
+ name: "T2"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_QINT8
+ type: DT_QUINT8
+ type: DT_QINT32
+ type: DT_QINT16
+ type: DT_QUINT16
+ }
+ }
+ }
+ attr {
+ name: "Tbias"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_FLOAT
+ type: DT_QINT32
+ }
+ }
+ }
+ attr {
+ name: "Toutput"
+ type: "type"
+ allowed_values {
+ list {
+ type: DT_FLOAT
+ }
+ }
+ }
+ attr {
+ name: "transpose_a"
+ type: "bool"
+ default_value {
+ b: false
+ }
+ }
+ attr {
+ name: "transpose_b"
+ type: "bool"
+ default_value {
+ b: false
+ }
+ }
+ attr {
+ name: "input_quant_mode"
+ type: "string"
+ default_value {
+ s: "MIN_FIRST"
+ }
+ allowed_values {
+ list {
+ s: "MIN_FIRST"
+ s: "SCALED"
+ }
+ }
+ }
+}
+op {
name: "QuantizedMatMulWithBiasAndRelu"
input_arg {
name: "a"