diditforlulz273 / pokerrl-omaha Goto Github PK
View Code? Open in Web Editor NEWOmaha Poker functionality+some features for PokerRL Reinforcement Learning card framwork
License: MIT License
Omaha Poker functionality+some features for PokerRL Reinforcement Learning card framwork
License: MIT License
You can delete me. I am just saying thanks for your contributions
As strategies buffer grows, memory usage grows rapidly, because every strategy copy its own lookup table inside its network. In my situation, memory grows out of memory at iteration 300+...
Every lookup table in the strategy are transferred into float32 occupying more than 800mb memory which make model training and serving impossible.
I was wondering, does this for allow for stack sizes to be added as an input parameter too the network and if so how long would you think it would take to converge to something close enough to the Nash equilibrium with respect to the normal un-ajusted stacksize version of this fork?
Also how would I go about adding stack size as an input parameter to this fork?
Hi Vsevolod!
I've tried to launch PLO_training_start.py with enabled LBR and failed (without any eval_methods iterations are running fine, but I can't evaluate results). I've tried both PLO and DiscretizedNLHoldem, with Debugging option turned on and off.
When DEBUGGING=True, and nn_type "feedforward" or "dense_residual", I've got AssertionError:
/PokerRL-Omaha-master/DeepCFR/IterationStrategy.py", line 144, in get_a_probs_for_each_hand_in_list
assert len(pub_obs.shape) == 2, "all hands have the same public obs"
AssertionError: all hands have the same public obs
And if DEBUGGING=False I've got this error on iteration 1:
PokerRL-Omaha-master/PokerRL/rl/neural/MainPokerModuleFLAT2.py", line 109, in forward
pf_mask = torch.where(pub_obses[:, 14] == 1)
TypeError: list indices must be integers or slices, not tuple
If nn_type="recurrent", I've got error on iteration 0:
PokerRL-Omaha-master/PokerRL/rl/neural/MainPokerModuleRNN.py", line 157, in forward
pub_obses = torch.from_numpy(pub_obses[0]).to(self.device).view(seq_len, bs, self.pub_obs_size)
TypeError: expected np.ndarray (got Tensor)
My requirements.txt:
gym==0.10.9 (tried 0.12.5 too)
numpy==1.21.2
psutil==5.8.0
pycrayon==0.5
pytz==2021.3
ray==0.6.1 (didn't use Distributed)
scipy==1.7.3
torch==1.4.0 (tried Pytorch versions till 1.10 with CUDA 10.2)
Where is the code for the CPU:-GPU training scheme? for me it is not using the GPU
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.