Search

Browse

My Account

Discover

  • Author

  • Subject

    • Applied mathematics (1)
    • Electronic Discovery, Markov Decision Process (MDP), Multi-Armed Bandit (MAB), Optimization Under Uncertainties, Sampling, Stochastic Scheduling (1)
    • ... View More
  • Date Issued

    • 2014 (1)