You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-transform-data.md
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -285,7 +285,7 @@ dataflow.head(5)
285
285
286
286
### Filtering columns
287
287
288
-
To filter columns, use `Dataflow.drop_columns()`. This method takes a list of columns to drop or a more complex argument called [`ColumnSelector`](https://docs.microsoft.com/en-us/python/api/azureml-dataprep/azureml.dataprep.columnselector?view=azure-dataprep-py).
288
+
To filter columns, use `Dataflow.drop_columns()`. This method takes a list of columns to drop or a more complex argument called [`ColumnSelector`](https://docs.microsoft.com/python/api/azureml-dataprep/azureml.dataprep.columnselector?view=azure-dataprep-py).
289
289
290
290
#### Filtering columns with list of strings
291
291
@@ -490,7 +490,7 @@ df.head(2)
490
490
|0|ALABAMA|Jefferson County|Jefferson County, Alabama|1.019200e+10|1.0|
491
491
|1|ALABAMA|Jefferson County|Jefferson County, Alabama|1.019200e+10|0.0|
492
492
493
-
## Next Steps
493
+
## Next steps
494
494
495
495
* See the SDK [overview](https://aka.ms/data-prep-sdk) for design patterns and usage examples
496
496
* See the Azure Machine Learning Data Prep SDK [tutorial](tutorial-data-prep.md) for an example of solving a specific scenario
Copy file name to clipboardExpand all lines: articles/machine-learning/service/how-to-write-data.md
+27-41Lines changed: 27 additions & 41 deletions
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ ms.custom: seodec18
15
15
---
16
16
# Write data using the Azure Machine Learning Data Prep SDK
17
17
18
-
In this article, you learn different methods to write data using the Azure Machine Learning Data Prep SDK. Output data can be written at any point in a dataflow, and writes are added as steps to the resulting data flow and are run every time the data flow is. Data is written to multiple partition files to allow parallel writes.
18
+
In this article, you learn different methods to write data using the [Azure Machine Learning Data Prep SDK](https://aka.ms/data-prep-sdk). Output data can be written at any point in a dataflow, and writes are added as steps to the resulting data flow and are run every time the data flow is. Data is written to multiple partition files to allow parallel writes.
19
19
20
20
Since there are no limitations to how many write steps there are in a pipeline, you can easily add additional write steps to get intermediate results for troubleshooting or for other pipelines.
21
21
@@ -27,7 +27,7 @@ The following file formats are supported
27
27
- Delimited files (CSV, TSV, etc.)
28
28
- Parquet files
29
29
30
-
Using the [Azure Machine Learning Data Prep python SDK](https://aka.ms/data-prep-sdk), you can write data to:
30
+
Using the Azure Machine Learning Data Prep python SDK, you can write data to:
31
31
+ a local file system
32
32
+ Azure Blob Storage
33
33
+ Azure Data Lake Storage
@@ -48,22 +48,17 @@ For this example, start by loading data into a data flow. You reuse this data wi
48
48
import azureml.dataprep as dprep
49
49
t = dprep.auto_read_file('./data/fixed_width_file.txt')
|3| 10013.0 | 99999.0 | ERROR | NO | NO || NaN | NaN | NaN |
85
+
|4| 10014.0 | 99999.0 | ERROR | NO | NO | ENSO | 59783.0 | 5350.0 | 500.0|
86
+
96
87
97
88
In the preceding output, several errors appear in the numeric columns because of numbers that were not parsed correctly. When written to CSV, null values are replaced with the string "ERROR" by default.
0 commit comments