Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Status of testing Providers that were prepared on May 19, 2023 #31322

Closed
33 of 80 tasks
eladkal opened this issue May 16, 2023 · 24 comments
Closed
33 of 80 tasks

Status of testing Providers that were prepared on May 19, 2023 #31322

eladkal opened this issue May 16, 2023 · 24 comments
Labels
kind:meta High-level information important to the community testing status Status of testing releases

Comments

@eladkal
Copy link
Contributor

eladkal commented May 16, 2023

I have a kind request for all the contributors to the latest provider packages release.
Could you please help us to test the RC versions of the providers?

Let us know in the comment, whether the issue is addressed.

Those are providers that require testing as there were some substantial changes introduced:

Provider amazon: 8.1.0rc2

Provider apache.beam: 5.1.0rc2

Provider apache.hdfs: 4.0.0rc2

Provider apache.hive: 6.1.0rc2

Provider apache.pinot: 4.1.0rc2

Provider cncf.kubernetes: 7.0.0rc2

Provider common.sql: 1.5.0rc2

Provider databricks: 4.2.0rc2

Provider dbt.cloud: 3.2.0rc2

Provider elasticsearch: 4.5.0rc2

Provider google: 10.1.0rc2

Provider microsoft.azure: 6.1.0rc2

Provider mongo: 3.2.0rc2

Provider neo4j: 3.3.0rc2

Provider oracle: 3.7.0rc2

Provider pagerduty: 3.2.0rc2

Provider redis: 3.2.0rc2

Provider slack: 7.3.0rc2

The guidelines on how to test providers can be found in

Verify providers by contributors

All users involved in the PRs:
@ying-w @alextgu @RachitSharma2001 @hussein-awala @jonshea @vincbeck @dacort @VVildVVolf @pankajastro @shahar1 @potiuk @vandonr-amz @ferruzzi @eladkal @utkarsharma2 @harishkesavarao @bkossakowska
@dstandish @jon-evergreen @tnk-ysk @IAL32 @nsAstro @Owen-CH-Leung @eldar-eln-bigabid @ephraimbuddy @moiseenkov @jbbqqf @attilaszombati @ahidalgob @vchiapaikeo @JCoder01 @syedahsn @pankajkoti @lwyszomi
@phanikumv

@eladkal eladkal added kind:meta High-level information important to the community testing status Status of testing releases labels May 16, 2023
@dacort
Copy link
Contributor

dacort commented May 16, 2023

Confirmed #31169

@shahar1
Copy link
Contributor

shahar1 commented May 16, 2023

@raphaelauv
Copy link
Contributor

raphaelauv commented May 17, 2023

Provider cncf.kubernetes: 6.2.0rc1
add missing read for K8S config file from conn in deferred KubernetesPodOperator
✔️ #29498:

@harishkesavarao
Copy link
Contributor

Provider databricks: 4.2.0rc1

@ying-w
Copy link
Contributor

ying-w commented May 17, 2023

Confirmed #30655 by verifying priority on ui when running the following dag

from pendulum import datetime
from airflow.providers.common.sql.operators.sql import SQLExecuteQueryOperator
from airflow.decorators import task, dag

default_args = {
    "owner": "example",
    "retries": 0,
}

@dag(
    description="tmp",
    start_date=datetime(2023, 5, 1),
    schedule=None,
)
def testing():
    SQLExecuteQueryOperator(
        task_id="test",
        conn_id="bigquery",
        sql="SELECT 2",
        hook_params={
            "use_legacy_sql": False,
            "location": "us",
            "priority": "BATCH",  # this is the new part
            "api_resource_configs": {"query": {"useQueryCache": False}},
        },
    )

testing()

@phanikumv
Copy link
Contributor

#30829: works good. Tested it by passing a shareIdentifier while submitting the Batch job

[2023-05-17, 08:57:54 UTC] {batch_client.py:328} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) check status (RUNNABLE) in ['RUNNING', 'SUCCEEDED', 'FAILED']
[2023-05-17, 08:57:54 UTC] {batch_client.py:343} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) status check (7 of 4200) in the next 7.67 seconds
[2023-05-17, 08:58:03 UTC] {batch_client.py:328} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) check status (STARTING) in ['RUNNING', 'SUCCEEDED', 'FAILED']
[2023-05-17, 08:58:03 UTC] {batch_client.py:343} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) status check (8 of 4200) in the next 8.68 seconds
[2023-05-17, 08:58:12 UTC] {batch_client.py:328} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) check status (SUCCEEDED) in ['RUNNING', 'SUCCEEDED', 'FAILED']
[2023-05-17, 08:58:21 UTC] {batch_client.py:328} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) check status (SUCCEEDED) in ['SUCCEEDED', 'FAILED']
[2023-05-17, 08:58:21 UTC] {batch_client.py:269} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) has completed
[2023-05-17, 08:58:21 UTC] {batch.py:310} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) CloudWatch Events details found. Links to logs:
[2023-05-17, 08:58:21 UTC] {batch.py:313} INFO - 
[2023-05-17, 08:58:21 UTC] {batch_client.py:245} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) succeeded: {'jobArn': 'arn:aws:batch:us-east-2:396807896376:job/733b153e-fda5-4f3c-a29b-95eb8c22a6d4', 'jobName': 'test_***_job', 'jobId': '733b153e-fda5-4f3c-a29b-95eb8c22a6d4', 'jobQueue': 'arn:aws:batch:us-east-2:396807896376:job-queue/providers_team_job_queue', 'status': 'SUCCEEDED', 'shareIdentifier': 'providers', 'schedulingPriority': 1, 'attempts': [{'container': {'containerInstanceArn': 'arn:aws:ecs:us-east-2:396807896376:container-instance/AWSBatch-providers_team_compute_env-0f056134-f57a-3a1f-9cbb-8d84bdd0f935/bcefcb90921e460085877d330bc5f495', 'taskArn': 'arn:aws:ecs:us-east-2:396807896376:task/AWSBatch-providers_team_compute_env-0f056134-f57a-3a1f-9cbb-8d84bdd0f935/0c4cdbb8cbfa4605aa1ad82e28384f87', 'exitCode': 0, 'logStreamName': 'providers_team_job_defn/default/0c4cdbb8cbfa4605aa1ad82e28384f87', 'networkInterfaces': []}, 'startedAt': 1684313885354, 'stoppedAt': 1684313885432, 'statusReason': 'Essential container in task exited'}], 'statusReason': 'Essential container in task exited', 'createdAt': 1684313824707, 'startedAt': 1684313885354, 'stoppedAt': 1684313885432, 'dependsOn': [], 'jobDefinition': 'arn:aws:batch:us-east-2:396807896376:job-definition/providers_team_job_defn:2', 'parameters': {}, 'container': {'image': 'public.ecr.aws/amazonlinux/amazonlinux:latest', 'command': ['echo', 'hello world'], 'volumes': [], 'environment': [], 'mountPoints': [], 'ulimits': [], 'exitCode': 0, 'containerInstanceArn': 'arn:aws:ecs:us-east-2:396807896376:container-instance/AWSBatch-providers_team_compute_env-0f056134-f57a-3a1f-9cbb-8d84bdd0f935/bcefcb90921e460085877d330bc5f495', 'taskArn': 'arn:aws:ecs:us-east-2:396807896376:task/AWSBatch-providers_team_compute_env-0f056134-f57a-3a1f-9cbb-8d84bdd0f935/0c4cdbb8cbfa4605aa1ad82e28384f87', 'logStreamName': 'providers_team_job_defn/default/0c4cdbb8cbfa4605aa1ad82e28384f87', 'networkInterfaces': [], 'resourceRequirements': [{'value': '1', 'type': 'VCPU'}, {'value': '2048', 'type': 'MEMORY'}], 'secrets': []}, 'tags': {}, 'propagateTags': False, 'platformCapabilities': ['EC2'], 'eksAttempts': []}
[2023-05-17, 08:58:21 UTC] {batch.py:329} INFO - AWS Batch job (733b153e-fda5-4f3c-a29b-95eb8c22a6d4) succeeded

@tnk-ysk
Copy link
Contributor

tnk-ysk commented May 17, 2023

I tested #30516, it works as expected

@phanikumv
Copy link
Contributor

#30968 works as expected

[2023-05-17, 10:13:51 UTC] {dbt.py:488} INFO - Getting the status of job run 151181329.
[2023-05-17, 10:13:51 UTC] {base.py:73} INFO - Using connection ID 'dbt_test_conn' for task execution.
[2023-05-17, 10:13:52 UTC] {dbt.py:493} INFO - Current status of job run 151181329: SUCCESS
[2023-05-17, 10:13:52 UTC] {taskinstance.py:1345} INFO - Marking task as SUCCESS. dag_id=example_dbt_cloud, task_id=job_run_sensor_async, execution_date=20230517T101254, start_date=20230517T101351, end_date=20230517T101352
[2023-05-17, 10:13:52 UTC] {local_task_job_runner.py:225} INFO - Task exited with return code 0

@phanikumv
Copy link
Contributor

#30579, #30939, #30618, #30920, #30901 work as expected

@potiuk
Copy link
Member

potiuk commented May 17, 2023

Checked and marked all mine as ✅ (except #31080 - which looks good but I have no databricks account/config to check it with) . Looks good. @Stormhand I would appreciate if you could check if #31080 is solved for you - you'd need to install https://pypi.org/project/apache-airflow-providers-databricks/4.2.0rc1 (it should pull common-sql 1.5.0rc1 automatically).

@pankajkoti
Copy link
Member

Confirmed #31063, #31042, and #31062 (all of them are for Provider amazon: 8.1.0rc1) are included in the RC and they ran fine on my example DAG. Thank you for the efforts!

@utkarsharma2
Copy link
Contributor

utkarsharma2 commented May 18, 2023 via email

@JCoder01
Copy link
Contributor

#31110 works as expected

@vandonr-amz
Copy link
Contributor

👍 for me

@eladkal eladkal changed the title Status of testing Providers that were prepared on May 16, 2023 Status of testing Providers that were prepared on May 19, 2023 May 19, 2023
@eladkal
Copy link
Contributor Author

eladkal commented May 19, 2023

Issue template has been updated for RC2

@hussein-awala
Copy link
Member

Should we create a major release for google provider? #30067 worth it. @eladkal WDYT?

@eladkal
Copy link
Contributor Author

eladkal commented May 20, 2023

Should we create a major release for google provider? #30067 worth it. @eladkal WDYT?

I didnt see breaking changes on that PR. Can you specify which change?
Depended packages breaking changes are not considered breaking change for Airflow. We do major release when Airflow breaks the schema/signature of functions/classes/imports.

@potiuk
Copy link
Member

potiuk commented May 20, 2023

Should we create a major release for google provider? #30067 worth it. @eladkal WDYT?

I didnt see breaking changes on that PR. Can you specify which change? Depended packages breaking changes are not considered breaking change for Airflow. We do major release when Airflow breaks the schema/signature of functions/classes/imports.

That's right. We even documented it: https://github.com/apache/airflow#semantic-versioning

Changing limits for versions of Airflow dependencies is not a breaking change on its own.

If there were no breaking changes in OUR API in providers, deps upgrade are not breaking. This is quite common approach - for example apache-beam uses the same. The version that enabled the upgrade was 2.47.0, a minor upgrade from previous 2.46.0 - and they bumped some of the same deps we did (pyarrow, protobuf to name the most important ones).

@potiuk
Copy link
Member

potiuk commented May 21, 2023

Everything Looks good:

  • no more circular import problems (I checked it in similar way as is done in our CI now: Check Logging Handlers and Secret Backends for recursion #31385 - but I did it also for Airflow 2.5 and 2.6.1 (not only 2.4 and main).
  • All providers now have __version__ defined at their top-level
  • I also checked that this version of preinstalled providers cannot be used for airflow < 2.4. When you try to import those you get this nice exception:

image

@potiuk potiuk mentioned this issue May 21, 2023
11 tasks
@pankajastro
Copy link
Member

#31006 #31094 #30744 #30761 looks good. thank you!

@Lee-W
Copy link
Member

Lee-W commented May 22, 2023

I tested the following DAG

    create_job_flow = EmrCreateJobFlowOperator(
        task_id="create_job_flow",
        job_flow_overrides=JOB_FLOW_OVERRIDES,
        aws_conn_id=AWS_CONN_ID,
    )

but encountered the following issue

Traceback (most recent call last):
  File "/usr/local/lib/python3.10/site-packages/airflow/providers/amazon/aws/operators/emr.py", line 695, in execute
    log_uri=get_log_uri(emr_client=self._emr_hook.conn, job_flow_id=self._job_flow_id),
  File "/usr/local/lib/python3.10/site-packages/airflow/providers/amazon/aws/links/emr.py", line 61, in get_log_uri
    log_uri = S3Hook.parse_s3_url(response["Cluster"]["LogUri"])
KeyError: 'LogUri'

@potiuk
Copy link
Member

potiuk commented May 22, 2023

Looks like might be caused by #31032 (cc: @ferruzzi )

@phanikumv
Copy link
Contributor

phanikumv commented May 22, 2023

I tested #30252 and found a bug in WasbPrefixSensor while testing microsoft.azure rc, fix is available in this PR #31411 . Would it be okay to include it as well

@eladkal
Copy link
Contributor Author

eladkal commented May 22, 2023

Thank you everyone.
Providers are released
I invite everyone to help improve providers for the next release, a list of open issues can be found here.

Note: we decided to release amazon and microsoft.azure providers despite the bug reports, reasons are explained in the mailing list thread.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind:meta High-level information important to the community testing status Status of testing releases
Projects
None yet
Development

No branches or pull requests