@@ -256,6 +256,7 @@ def test_remote_runner_translation(self):
256
256
def test_streaming_create_translation (self ):
257
257
remote_runner = DataflowRunner ()
258
258
self .default_properties .append ("--streaming" )
259
+ self .default_properties .append ("--experiments=disable_runner_v2" )
259
260
with Pipeline (remote_runner , PipelineOptions (self .default_properties )) as p :
260
261
p | ptransform .Create ([1 ]) # pylint: disable=expression-not-assigned
261
262
job_dict = json .loads (str (remote_runner .job ))
@@ -838,15 +839,21 @@ def test_group_into_batches_translation_non_unified_worker(self):
838
839
'Runner determined sharding not available in Dataflow for '
839
840
'GroupIntoBatches for jobs not using Runner V2' ):
840
841
_ = self ._run_group_into_batches_and_get_step_properties (
841
- True , ['--enable_streaming_engine' ])
842
+ True ,
843
+ ['--enable_streaming_engine' , '--experiments=disable_runner_v2' ])
842
844
843
845
# JRH
844
846
with self .assertRaisesRegex (
845
847
ValueError ,
846
848
'Runner determined sharding not available in Dataflow for '
847
849
'GroupIntoBatches for jobs not using Runner V2' ):
848
850
_ = self ._run_group_into_batches_and_get_step_properties (
849
- True , ['--enable_streaming_engine' , '--experiments=beam_fn_api' ])
851
+ True ,
852
+ [
853
+ '--enable_streaming_engine' ,
854
+ '--experiments=beam_fn_api' ,
855
+ '--experiments=disable_runner_v2'
856
+ ])
850
857
851
858
def test_pack_combiners (self ):
852
859
class PackableCombines (beam .PTransform ):
0 commit comments