Requesting for Adminstration

Jump to navigation Jump to search

Requesting for Adminstration

I would like to be a Administrator here for having access to Special:ReplaceText, I think after near one year activity here and near 2k edit I have enough experience for using this extension wisely. I don't know granting access in translatewiki is community based or not but I think the Persian translators community can trust me for having this access. Special:ReplaceText is very good for finding exact instance of one Regex template in texts and I think I know needed Regex sufficient.

ebraminiotalk22:50, 28 March 2011
granted granted
Mardetanha talk23:03, 28 March 2011

Sorry to step in here. ReplaceText is not to be used by users with the Administrator/sysop role; we reserve usage of this functionality to staff. I am removing this right for administrators. We suggest using a pywikipediabot.

Siebrand08:53, 29 March 2011

You do not have permission to edit this page, for the following reason:

The action you have requested is limited to users in the group: Users.

You can view and copy the source of this page.

Return to Thread:Project:Translator/Requesting for Adminstration/reply (3).



ebraminiotalk18:24, 9 April 2011

Please explain your problem. I can provide a list of pages in a particular language.

Siebrand19:09, 9 April 2011

This is good, but the thing that I really need is having a tool for finding occurrence of a regex (and chars) on that pages. I know with list of my language translations I can do that with pywikipedia (I used pywikipedia in my wikipedia for a lot propuses). But this can done better and easier without that.

For example we have some hidden character in Persian (zwnj, zwj and direction control chars that is common in RTL languages). If I had a tool for finding misused occurrence of them in translation, I could easily fix them. For example this regex needed for my language (and some languages with Arabic script): s/([ازرذدژوة])\u200c/\1/g also finding non-standard chars (for my languages) with Special:Search is not possible because Special:Search only yields occurrence of words and not chars.

ebraminiotalk19:57, 9 April 2011

Using regex on over a million pages in will be blocking our database for too long. We really cannot make that functionality available. If you were to provide the regexes you are looking for, we will try and find a way to help you out.

Siebrand23:49, 9 April 2011
Edited by author.
Last edit: 11:11, 11 April 2011

Okay, Thanks for your great jobs on TranslateWiki. Seems i must use Pywikipedia. If you have time, Please make list of Persian translations for me. Also SQL or XML dump of this wiki (separated for each language) or a replication on toolserver is other solutions I think.

May it is better this thread moved to Portal:Fa. Thanks again!

ebraminiotalk08:37, 11 April 2011

Feel free to move the thread. Will get back with a reply. Data dump: have to ask Niklas.

Siebrand09:28, 11 April 2011

Okay, Thanks :)

ebraminiotalk11:14, 11 April 2011

I'm running a backup script now. We'll have to inspect what this results in. If it's OK, we can schedule a weekly dump.

php ./dumpBackup.php --current --report=1000 --skipheader --skip-footer --conf=../LocalSettings.php --output=bzip2:$HOME/dump.bzip2

Siebrand12:44, 13 April 2011

Backup is ready. Took about 20 minutes to create. Now checking if there's not something in there we wouldn't want to distribute (I think that there are no issues, but better safe than sorry). Will now also prepare a list of '/fa' pages with namespace.

Siebrand13:29, 13 April 2011


  • fa.txt contains all pages ending "/fa"
  • contains an XML dump of
Siebrand11:31, 15 April 2011

Thank you. I don't need fa.txt because I will use "LIKE" on titles (also this is not needed really for me), But just a little problem, I can not download huge files in periods of time because it will consume my Internet traffic limits and my Internet traffic is so limited.

Can you split this dumps if you have program for publishing this dump also in future?

ebraminiotalk19:30, 15 April 2011

Split what/how? It's "only" a 77MB download.

Siebrand21:17, 15 April 2011

Splitting dump of each language. This tool is useful if you want.

Sadly, yes, it is only "77mb" but is big for me :(

ebraminiotalk22:22, 15 April 2011

Download it patiently then!

Huji15:04, 17 April 2011

Okay, I will solve my problem myself :) Thanks for your great helps.

ebraminiotalk17:27, 17 April 2011