I have a pyspark dataframe
| id | events |
|---|---|
| a0 | a-markets-l1 |
| a0 | a-markets-watch |
| a0 | a-markets-buy |
| c7 | a-markets-z2 |
| c7 | scroll_down |
| a0 | a-markets-sell |
| b2 | next_screen |
I am trying to join events by grouping IDs Here's my python code
df_events_userpath = df_events.groupby('id').agg({ 'events': lambda x: ' '.join(x)}).reset_index()
| id | events |
|---|---|
| a0 | a-markets-l1 a-markets-watch a-markets-buy a-markets-sell |
| c7 | a-markets-z2 scroll_down |
| b2 | next_screen |