There are some hidden characters inside like "়া" (without the last one) which are counted with len() but I have 5 characters available until the rigtht side of a 80 column page. Maybe this is a such a minor problem that we just could decline it. I guess it does not worth to fix it.
Fri, Aug 17
We have moved to git and are no longer on svn. Can you show us your output of python pwb.oy version?
Wed, Aug 15
Fri, Aug 10
Tue, Aug 7
Fri, Aug 3
py 3.3 is no longer supported by pwb
See comments above
Sun, Jul 29
Decoding sys.argv would solve this issue above
absolute_path = abspath(os.path.dirname(sys.argv.decode('utf-8')))
but the same problem occurs when importing:
This is a core 2.0 issue which is no longer maintained.
Could you explain a bit please.
Is there a reason to leave her?
I guess the manual at mediawiki is a better place for the manual than docs because it is easier to update and it sin not needed to submit changes throught the code review workflow. Whereas docs is the right place for the api documentation and additional stuff related to it. I tend to remove that warning banner.
Wed, Jul 25
See also T130911
I've added the L10N for content dict to use clean_sandbox for "bn" site code. This works for all projects wich have a sandbox page listed on wikidata item Q3938. For he others the -page option has to be used to choose the page. Or create a wikidata item for it which could be added to the "sandbox_titles" item tuple.
Tue, Jul 24
-easy: this is not trivial. First we need the common content for the sandbox page but what is it? Either one for all or one for each site or one for each page. Currently 'bn' isn't localized. Second the default project page is either Q3938 or Q28939665. Now we have 3 pages for each 'bn' site. Finally multipe -page options may work but the content is still missing and -text option is needed.
Mon, Jul 23
Sun, Jul 22
This is valid for compat only
Jul 17 2018
I'm confused, there already are some deprecations in 7 scripts. So we just need to start deprecating in scripts? I would support to finish https://gerrit.wikimedia.org/r/#/c/pywikibot/core/+/399755/ ASAP for this.
There are two L10N dicts in commonscat already. What is needed further?
Maybe we should move scripts directory to _scripts to make this more explicit?
Can somebody update what is to be updated here please.
Jul 16 2018
I also cannot follow this request. Could you explain please or make an example of what do you expect and what is the current behavior.
Is that implementable?
bot and redirect parameters are already implemented for newpages. Adding this to NewpagesPageGenerator does not make any sense because GeneratorFactory.handleArg cannot retrieve these parameters.
See that sample given above:
Ah, I see: The problem is the check inside initializer.
Jul 15 2018
Cannot follow this task. Pywikibot is throttleling not to edit more than 6 (I guess) pages per minute by default. What might be the reason that the limit of 40 pages is exceeded. What was the putthrottle setting when the exception occurred?
It can have different settings already. Just add it to the xdict L10N table „content“ like:
I did some measurements for the memory needed for the statements given above:
Jul 14 2018
- use hash function for the filter_unique key
- use hash function for the filter_unique key by default
- use a GeneratorFactory Container attribute to hold the seen pages which could be reused when we have more than one duplicate filter
- use an container which uses disk space instead of memory (but this could be time consuming)
No glue where the Memory leakage might come from.
I see the getsizeof() counts the pointers only but not the Page objects itself.
Long-running tasks may end on MemoryError to filter_unique leaks Memory
Why do you assume that?
This is too undetermined. Could you give a sample?
When filter_unique is used the filter key could be replaced by the hash but this does not solve a lot. Most space is used by the content but it is not clear to me where this is hold after processing.
unsigned comments are valid threads and archivebot ignores them. Unfortunately these threads leads test to fail but we must not ignore them otherwise we cannot check whether the regexes work properly. How proceed in this case?
- either add a timestamp to the talk thread
- or use an alternate talk page for a given site
All other exceptions means there is an undetected timestamp and should be fixed be the regex.
The error message is equal to that in T69663. Can we solve it at once?
Jul 13 2018
Jul 11 2018
Jul 10 2018
No glue why this rdirection change was introduced with T129193; I propose to revert it as suggested.
Seems it is solved upstream. No further failure seen in tests.
Jul 9 2018
Seems this edit shouldn't have been done too.