@@ -398,7 +398,7 @@ def _test_with_attention(
398398 seed = 1337
399399 )
400400
401- policy = tf .keras .mixed_precision .experimental . global_policy ()
401+ policy = tf .keras .mixed_precision .global_policy ()
402402 sampler = sampler_py .TrainingSampler ()
403403 my_decoder = basic_decoder .BasicDecoder (cell = cell , sampler = sampler )
404404 initial_state = cell .get_initial_state (
@@ -541,7 +541,7 @@ def set_random_state_for_tf_and_np():
541541@pytest .mark .usefixtures ("run_with_mixed_precision_policy" )
542542def test_bahdanau_not_normalized ():
543543 set_random_state_for_tf_and_np ()
544- policy = tf .keras .mixed_precision .experimental . global_policy ()
544+ policy = tf .keras .mixed_precision .global_policy ()
545545 create_attention_mechanism = wrapper .BahdanauAttention
546546 create_attention_kwargs = {"kernel_initializer" : "ones" }
547547 expected_final_output = basic_decoder .BasicDecoderOutput (
@@ -617,7 +617,7 @@ def test_bahdanau_normalized():
617617@pytest .mark .usefixtures ("run_with_mixed_precision_policy" )
618618def test_luong_not_normalized ():
619619 set_random_state_for_tf_and_np ()
620- policy = tf .keras .mixed_precision .experimental . global_policy ()
620+ policy = tf .keras .mixed_precision .global_policy ()
621621 create_attention_mechanism = wrapper .LuongAttention
622622
623623 expected_final_output = basic_decoder .BasicDecoderOutput (
0 commit comments