User Details
- User Since
- Mar 29 2021, 8:46 PM (245 w, 18 h)
- Availability
- Available
- LDAP User
- Unknown
- MediaWiki User
- Rachita Saha [ Global Accounts ]
May 3 2021
I have had an amazing experience working on this task and I truly appreciate how the community came forward to help out in any way possible everytime I or anyone else faced an issue. I have learnt a lot under the guidance of the mentors. I hope to apply all the knowledge I have gained through this experience in all future projects that I take up. I am really looking forward to continue contributing to open source.
Apr 29 2021
Apr 26 2021
Apr 24 2021
Apr 17 2021
Hello @rachita_saha it's true that we would face a ValueError in just one of the lines but is it actually fine to ignore it? On examining that particular row we would see that it has the data of a lot(really a lot) of different rows all combined to one probably due to some error by us in parsing (on the occurance of some special character sequence)
Apr 3 2021
Apr 2 2021
@Isaac Okay, understood. Thank you.
@Isaac @MGerlach I had a doubt regarding recording contributions. Since we are not making formal pull requests in this project, at what points and in which form are we required to record our contributions? Do we need to submit the public link of our notebook after completing a few to-dos and that will count as a contribution?
Apr 1 2021
Thank you @Isaac, I'll take these points into consideration.
Mar 31 2021
@MGerlach @Isaac Several of the popular destination articles do not link to any further articles. So there is no data on common pathways from this article. Should we find an article which links to more articles or should we only provide the visualizations for the common pathways to the chosen article?
Hello @Tru2198, I faced the same problem while converting the file to CSV. There were about 5-7 lines in the whole dataframe which had a variation in number of columns, probably due to error in data entry. I believe that since the number of bad lines is negligible as compared to the total number of lines, the parameter error_bad_lines = False can be used. However, @Isaac clarified that we can limit the data to something more feasible (20-30k) rows for this particular task. I did not find any erroneous lines in the first 20k entries so you could probably try working with a limited number too first.
Mar 30 2021
@Isaac Understood. Thank you for the clarification.
