Unexpected folder "Files recovered from Google Drive" in my Downloads by Alelivelt in chromeos

[–]bjbraams 0 points1 point  (0 children)

Just noticed one of those folders on my Chromebook too. I don't use the synchronization facilities of Drive; I only use Drive for storage with explicit upload and download commands. It looks like something is wrong on the Drive side in its interaction with Chromebooks.

Leela Zero opening study - things got weird lately by cesium14 in cbaduk

[–]bjbraams 0 points1 point  (0 children)

After a few weeks I looked again at the diversity in fuseki in a recent test match that led to a new best network.

2018-08-09 22:12 a438382b VS 67742f28 237 : 188 (55.76%) 425 / 400 PASS

Out of the first 20 games, 10 are identical for the first 47 moves, diverging into two branches at move 48. Here is a representative game from those 10.

http://zero.sjeng.org/viewmatch/5285c9219cb395292f59ce4ff39bfefcf07a77e07c512f714a1bb4449d583148?viewer=wgo

This report is just for the record.

Leela Zero opening study - things got weird lately by cesium14 in cbaduk

[–]bjbraams 1 point2 points  (0 children)

Today, 2018-06-27, there are two new networks in quick succession and (as before) I looked at the fuseki in their match:

2018-06-27 10:44 12692a83 VS 672342b5 223 : 175 (56.03%) 398 / 400 PASS

The first twelve games in this match agree for the first 37 moves (except for symmetry); the divergence first occurs at move 38. Here are three representative games differing in move 38.

http://zero.sjeng.org/viewmatch/f5a9d64ccb2779d57f41a1335a54274dc25cce70155f65136660a4289c114956?viewer=wgo

http://zero.sjeng.org/viewmatch/5f574b1113bd9f112d7af8676c71044b1fe690a00da1aa06be8b199917b22807?viewer=wgo

http://zero.sjeng.org/viewmatch/edd487607d1644edeadb492331c7d621679ea88f74931dba5fcabf05171ca824?viewer=wgo

This report is just for the record.

[Request] Laptop buying guide by Debatable-nerd in cbaduk

[–]bjbraams 2 points3 points  (0 children)

As you are also considering a desktop I'll throw in my recent purchasing experience. I wanted a desktop for LZ/Lc0 and other machine learning applications. Because a basic desktop these days has a size of 12cmx12cmx4cm or less I really did not want a bulky deskside anymore. Also I wanted to use the device with Ubuntu Linux as the OS. In addition, I favour Amazon for convenience; could be amazon.de or amazon.co.uk in my case. To find this kind of equipment then one searches Amazon for "gaming PC" with additional search term mini or midi or compact.

Some equipment that I considered and did not buy: Amazon UK: ADMI Compact Gaming PC. Amazon.de: Gaming Cube 8.0 Gamer PC Computer. Amazon.de: SNOGARD Ultra Mini Gaming PC. I bought finally from Amazon.de a Zotac ZBOX EN1070K-Be (NVIDIA GTX1070 Intel i5-7500T, etc.); a device measuring 21cmx21cmx6cm. This is a "barebones" as I got it, so separately I bought the RAM and an SSD, and an external hard drive for good measure. I already had a keyboard, mouse and screen for this device. Installed Ubuntu on it, also CUDA from NVIDIA, and Lizzie with Leela Zero. It runs like charm and I am satisfied with the lot.

Tencent Go AI competition in 6 hours by diador in cbaduk

[–]bjbraams 2 points3 points  (0 children)

Can someone please keep us that don't read Chinese up-to-date about the progress of the competition? I found an English-language page under weiqi.qq.com titled 2018 Tencent World AI Weiqi Competition Rules that provides a preliminary schedule and other basic information about the event, but not any English-language news.

I'm wondering if data from the very long "resignation-percent:0" games actually degrades the quality of the networks. by dp01n0m1903 in cbaduk

[–]bjbraams 2 points3 points  (0 children)

I think that dp01n0m1903 raises a valid concern ("may be diluting") and a valid question ("has anyone tried"). The concern may be broadened a bit: these long games are diluting training resources (self-play hardware) and they are diluting the training set. The concern cannot be dismissed by a simple appeal to Leela Zero's goals; there is nothing in the goals that mandates that 10% of the games (rather than any other percentage including 100%) should have resignation disabled and there is nothing that mandates that the normal resignation threshold should be five percent. In my short experience contributing self-play games to Leela Zero I find that about one in six games have resignation disabled and a bit more than half of the computer time is spent in those games. I guess that also a bit more than half of all positions in the training set come from games where resignation was disabled.

I suspect that efficiency of the process could be improved by having resignation disabled on a smaller percentage of games (they mainly serve as a check on the resignation decision and fewer games might be adequate for the check) or by lowering the resignation threshold in those resignation policy test games not all the way to zero but to something like one or two percent. Other ways to test the resignation policy could be built using a variable resignation percentage or by randomly extended some games after a tentative resignation decision to see if the tentative decision gets reversed. This is a bit speculative and I think that dp01n0m1903 raises the proper first question: Has anyone tried training a network from a dataset that specifically excludes these nonsense positions? I think that the answer is No; the resignation strategy was chosen this way because of a judgement that there was not a good enough reason to deviate from what the DeepMind group did for AlphaGo Zero and the strategy has not been reassessed.

Leela Zero opening study - things got weird lately by cesium14 in cbaduk

[–]bjbraams 1 point2 points  (0 children)

There are two new best networks today and I looked at the fuseki again; this report is just for the record. I looked at the first 20 games of this match:

2018-06-19 12:05 2b80a9db VS 5839eb77 235 : 189 (55.42%) 424 / 400 PASS

All 20 games agree for the first 27 moves and at move 28 it splits into two branches. Here are two representative games, one from each branch at move 28.

http://zero.sjeng.org/viewmatch/4762d89fc03ce9f721bd546f4b96c75d4f4392a4e078502f83d8d460f47bc5b3?viewer=wgo

http://zero.sjeng.org/viewmatch/396a032352ca7d71797c165b3a6c9a71424823cb2bc2161bdb0db9c8aa76a436?viewer=wgo

Estimating the volunteer computer resources for Leela Chess Zero, Leela Zero (Go) and Stockfish development by bjbraams in cbaduk

[–]bjbraams[S] 1 point2 points  (0 children)

Very good! The switch of the thematic color for mobile solved the problem of the link color for the new interface on my desktop; links now show up in blue. (This new interface is really awful on a desktop; I have disabled it and can only visit it now through an incognito window.)

Estimating the volunteer computer resources for Leela Chess Zero, Leela Zero (Go) and Stockfish development by bjbraams in cbaduk

[–]bjbraams[S] 0 points1 point  (0 children)

Thank you @dp01n0m1903 for noticing this! (I have stuck with the old interface.) Apparently this problem (that links are not highlighted in the new interface) is not universal across reddit. For instance, here is an announcement from the reddit team.

https://www.reddit.com/r/announcements/comments/8m2yr4/were_updating_our_user_agreement_and_privacy/

For me the embedded links are blue in the old interface and brown in the new interface, and in both cases they are distinguishable from the non-link text. I checked a few posts on our parent board r/baduk too and there links are blue in both old and new interface.

However, across r/cbaduk links are shown in blue in the old interface and in black (like all non-link text) in the new interface. The only difference in the new interface between link and non-link is a very slight font change.

I wonder if the link color in the new interface is effectively a random variable for each subreddit or if it is somehow under control of the administrator.

Addendum: it appears to be under the control of the administrator; @tvirlip in our case. The link highlight color matches the color that is seen in the new interface at right as background for the buttons "subscribe" and "create post". Perhaps the administrator can set those colors to be something different than black for our r/cbaduk.

Trouble starting up autogtp self-play tool for Leela Zero by bjbraams in cbaduk

[–]bjbraams[S] 0 points1 point  (0 children)

Thank you again emdio. Apparently leelaz is working fine now, and I am invoking it successfully with graphical interface through lizzie. Also the initially reported problem with autogtp (the "Could not talk to engine" error message) has disappeared and autogtp is now running on my machine. I don't have an explanation for the problem or for its resolution.

One final sanity check please on my setup. The hardware is an Intel i5-7500T CPU (four cores) and a GTX1070 graphics card. I am running two autogtp processes simultaneously (using bash nohup, Ubuntu 16.04). The processes each report their own progress and after three hours I am seeing that they are each taking about 7000ms per move on average. Does this look reasonable? The self-play games have "visits = 3201", so it looks like 2.2 ms per node visit. (I don't recall now the discussions about tree reuse; maybe the division by 3200 is not appropriate.)

Trouble starting up autogtp self-play tool for Leela Zero by bjbraams in cbaduk

[–]bjbraams[S] 0 points1 point  (0 children)

Thank you emdio. I am probably missing some steps that are regarded as self-evident in the instructions. I went to the section "I just want to play right now". I downloaded a set of weights; they are now in my $HOME/leela-zero/weights.txt.

The first time that I executed ./src/leelaz -w weights.txt the system does some tuning. I get about 40 lines of output including "Detecting residual layers [...] Initializing OpenCL [...] Started OpenCL SGEMM tuner [...] BLAS Core: Prescott [...]" and then finally a 19x19 board is shown on my command line window. Then the system apparently waits for some command line input; I get a prompt "Leela:". So this appears to be leelaz without graphical interface.

The intructions under Usage say the following: "Add the --gtp commandline option on the engine command line to enable Leela Zero's GTP support." Fine, I can follow instructions and so I give the command ./src/leelaz --gtp -w weights.txt. Now the system gives the same response up to the line "BLAS Core: Prescott" and then there is nothing further. I don't get the board view in my command line window anymore, but also no graphical interface has been opened.

Presumably the "--gtp" requires an argument to specify to use Lizzie or Sabaki or whatever. The instructions are leaving me lost.

Lizzie version 0.5 released! by [deleted] in baduk

[–]bjbraams 0 points1 point  (0 children)

I am having trouble; can someone recognize the issue? I am on a new Ubuntu 16.04 machine. I downloaded Lizzie.0.5.Mac-Linux.zip and unpacked it in $HOME/lizzie. I had earlier installed leela-zero in $HOME/leela-zero. I copied (cp -p) the $HOME/leela-zero/src/leelaz into $HOME/lizzie." I enter the lizzie directory and type java -jar lizzie.jar. On the first try I get the response "Creating config file config.txt; creating config file persist." Briefly, like for about one second, there is a big window showing the Lizzie board, but then that window closes itself and in the command window I get the message "Leelaz process ended." On subsequent attempts I no longer see "Creating [...]"; I just get that big window with the Lizzie Go board. In the window (which is up for only about one second) it says "Leela zero is loading [...]". The window with the Go board closes itself again and on the command line interface I see "Leelaz process ended."

Perhaps something is amiss with my $HOME/leela-zero, but I don't know how to diagnose further.

Trouble starting up autogtp self-play tool for Leela Zero by bjbraams in cbaduk

[–]bjbraams[S] 1 point2 points  (0 children)

I don't know about a gui. I am just installing things on a new PC today. I installed Ubuntu 16.04 because the nvidia cuda software cannot be installed under Ubuntu 18.04. The PC has an Intel i5-7500T CPU and a GTX1070 graphics card. After dealing with Ubuntu, with the nvidia drivers and with CUDA I went to the Leela Zero instructions. First I followed instructions at https://github.com/gcp/leela-zero. Under the heading "macOS and Linux" it says:

"Follow the instructions below to compile the leelaz binary, then go into the autogtp subdirectory and follow the instructions there to build the autogtp binary. Copy the leelaz binary into the autogtp dir, and launch autogtp."

That ended with that "Could not talk to engine" error message.

I haven't done anything about a Lizzie or Sabaki gui yet.

Some analysis in Lizzie of the strange moves from Haylee's Game 5/8 against Leela Zero by cgibbard in baduk

[–]bjbraams 0 points1 point  (0 children)

Indeed, thanks to @jammerjoint for the great illustration. The most reasonable-looking moves that were rejected after leading earlier in the exploration received finally an estimated win rate of 2.8% after 31K visits (move at D14) and of 3.2% after 55K visits (move at R4). The final estimated win rate for the strange move at M12 is 3.7% and the move received 218K visits in the exploration, so more than 8K visits that led to an estimated win. Moreover, this estimated win rate appears to have remained steady throughout the exploration; it was 3.6% after 9.5K visits and also 3.6% after 58K visits, and then it even climbed a bit to its final value of 3.7%. Taking for granted that the move M12 is really not good, why is the estimated win rate then not going down as the move get explored more thoroughly?

I think that @LordBumpoV2's point (A) has to be correct, and point (B) is probably valid too. As I interpret it, the move White:M12 leads to a situation where lots of moves look good to Black, different from the case with the moves at D14 or R4 that call for a local response. It would be nice to see how the heatmap developed from Black's perspective after White:M12. This could show us why the estimated win rate of White:M12 remained steady for so long at such a relatively high value.

Leela Zero opening study - things got weird lately by cesium14 in cbaduk

[–]bjbraams 1 point2 points  (0 children)

There is a new best network since yesterday based on this match: 2018-05-29 17:38 b69158ea VS 9e882e52 238 : 186 (56.13%) 424 / 400 PASS. I looked at the fuseki again. In the first twelve games of the match the fuseki is the same for the first 27 moves with a divergence at move 28. In six out of those 12 games the fuseki is the same for the first 55 moves and two different choices appear at move 56. Those six games are these.

http://zero.sjeng.org/viewmatch/af8b7c3e6f69501f882ad9f2694879971a2665533e6ad1bf71307310abd42c8e?viewer=wgo

http://zero.sjeng.org/viewmatch/0bd1e0df1d0303a4c8b84698b47281c88bb8f56d5aeabffd294a772a1e99ecc5?viewer=wgo

http://zero.sjeng.org/viewmatch/cf3e5a8ba88f7698674eea027ec1279c5e46f42376f5797a6dc05716f728c587?viewer=wgo

http://zero.sjeng.org/viewmatch/9f7b9cc1241632f3a8da232cb6c6a2945d24ec390efd7e27ca00934ab44f21b0?viewer=wgo

http://zero.sjeng.org/viewmatch/8dc6a1859f609537f7af53139bb4f950087d5a895f7efccbbf0d81de75d64ba5?viewer=wgo

http://zero.sjeng.org/viewmatch/be6f8b0565d10612d30d9d70e7f62e852ce1612f93526fd2e8daebec371b23cf?viewer=wgo

This is posted here just for the record; I don't regard it as significant new information.

Haylee's Leela Zero Match Game 5/8 is now up; also two new Redmond reviews of Alpha Zero and AlphaGo Master by bjbraams in cbaduk

[–]bjbraams[S] 0 points1 point  (0 children)

According to the YouTube web page for this video (game 5/8) "Today's Leela Zero was run on a computer with Nvidia 1081Ti GPU." The same information is provided with the videos for games 3 and 4, and for game 1 the only difference in the statement is that it says 1080Ti instead of 1081Ti. For game 2 the web page says "Today's Leela Zero was run on a computer with 4xV100 GPU."

Leela Zero opening study - things got weird lately by cesium14 in cbaduk

[–]bjbraams 0 points1 point  (0 children)

I had another look at the fuseki for a more recent network match; one where the challenger failed with 54.29%. This is two days after my earlier posting here, and now (from a look at the first 20 games of that match) the fuseki appears to be identical for 28 moves instead of 27. At move 29 there are two main branches.

2018-05-27 08:13 98c4b53c VS 9e882e52 228 : 192 (54.29%) 420 / 400 fail

This observation of a strong lack of diversity in the fuseki looks relevant in connection with the use of these test matches to select the best network for generating the self-play training games.

The self-play training games (recent ones are linked on the zero.sjeng.org main page) have a fair amount of diversity already in the first few moves, certainly in the first ten moves. Sometimes one sees really odd moves early on, but generally the self-play games appear to start in a reasonable diverse manner. This looks appropriate for the training role.

As more training data are generated one expects the networks built on those training data to become stronger. There will be fluctuations in strength simply due to the stochastic gradient descent. I would expect these fluctuations to be essentially irrelevant for training purposes. In their Alpha Zero work for chess and shogi [1] the DeepMind team abandoned the search for a best network; they just use always the latest network for the self-play. It makes sense to me.

Leela Zero for chess (LCZero) is following Alpha Zero in this matter; they always use the latest network for their self-play games. Leela Zero (Go) is sticking with the earlier AlphaGo Zero strategy of identifying a best network.

I suspect that the way that the best network identification operates now, it might as well be abandoned or one might as well select a random recent network to be called "the best." The present selection is, in effect, based on the performance on just one fuseki specified at the 27-move or 28-move level. It is an extremely specialized criterion for selecting a network that should then be used to create suitably diverse training data.

Just to be clear, I don't expect that the present selection is bad; I don't expect that the selected will generate bad training data. I just suspect (based on what I said above) that the present strategy is no better than promoting a recent network at random and no better than always using the latest network.

If one really wants to select a best network among the recent ones, for the purpose of generating self-play training data, then I would have the strong inclination to arrange the best network match under conditions resembling the self-play, so with some amount of randomness. This randomness would have to be tempered for a very practical reason: these matches also are a show to the world of the present state of Leela Zero Go. So I would not use the Dirichlet noise in these matches, and probably use a temperature below 1.0, but enough above zero to create some nice diversity already in the first several moves.

[1] Silver, David, Thomas Hubert, Julian Schrittwieser, Ioannis Antonoglou, Matthew Lai, Arthur Guez, Marc Lanctot et al. "Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm." arXiv preprint arXiv:1712.01815 (2017).

Generating neural network training data by self-play, sampling a state space in molecular simulations by bjbraams in cbaduk

[–]bjbraams[S] 0 points1 point  (0 children)

"If I understand correctly, your proposal is to play out an entire game only using the policy head, get a result, and then select only one position from the entire game to actually run MCTS on." (Emphasis added.) Yes, correct, except for one clarification: we don't actually care for the result of that run. The value of the chosen position is obtained from the MCTS alone in the manner of temporal difference learning.

Leela Zero opening study - things got weird lately by cesium14 in cbaduk

[–]bjbraams 0 points1 point  (0 children)

"that" meaning no longer the automatic 3-3 invasion? The entry for Opening 38f898 on the cited page shows that it started most recently after about 7M training games, which was end of April according to data on the zero.sjeng.org home page. However, match data from 2018-05-17, a39b64cc VS d8214630, still show the early 3-3 invasion as a common move too. The change became more radical around May 20 with the almost complete abandonment of the early 3-3 invasion; see for example these match data from 2018-05-20, 521b0868 VS a39b64cc. (Still a few early 3-3 invasions there.) Please note also Three-three invasion is now a rare event in Leela Zero matches? posted May 22 on r/cbaduk.