You can subscribe to this list here.
2003 |
Jan
|
Feb
|
Mar
|
Apr
|
May
(3) |
Jun
|
Jul
|
Aug
(12) |
Sep
(12) |
Oct
(56) |
Nov
(65) |
Dec
(37) |
---|---|---|---|---|---|---|---|---|---|---|---|---|
2004 |
Jan
(59) |
Feb
(78) |
Mar
(153) |
Apr
(205) |
May
(184) |
Jun
(123) |
Jul
(171) |
Aug
(156) |
Sep
(190) |
Oct
(120) |
Nov
(154) |
Dec
(223) |
2005 |
Jan
(184) |
Feb
(267) |
Mar
(214) |
Apr
(286) |
May
(320) |
Jun
(299) |
Jul
(348) |
Aug
(283) |
Sep
(355) |
Oct
(293) |
Nov
(232) |
Dec
(203) |
2006 |
Jan
(352) |
Feb
(358) |
Mar
(403) |
Apr
(313) |
May
(165) |
Jun
(281) |
Jul
(316) |
Aug
(228) |
Sep
(279) |
Oct
(243) |
Nov
(315) |
Dec
(345) |
2007 |
Jan
(260) |
Feb
(323) |
Mar
(340) |
Apr
(319) |
May
(290) |
Jun
(296) |
Jul
(221) |
Aug
(292) |
Sep
(242) |
Oct
(248) |
Nov
(242) |
Dec
(332) |
2008 |
Jan
(312) |
Feb
(359) |
Mar
(454) |
Apr
(287) |
May
(340) |
Jun
(450) |
Jul
(403) |
Aug
(324) |
Sep
(349) |
Oct
(385) |
Nov
(363) |
Dec
(437) |
2009 |
Jan
(500) |
Feb
(301) |
Mar
(409) |
Apr
(486) |
May
(545) |
Jun
(391) |
Jul
(518) |
Aug
(497) |
Sep
(492) |
Oct
(429) |
Nov
(357) |
Dec
(310) |
2010 |
Jan
(371) |
Feb
(657) |
Mar
(519) |
Apr
(432) |
May
(312) |
Jun
(416) |
Jul
(477) |
Aug
(386) |
Sep
(419) |
Oct
(435) |
Nov
(320) |
Dec
(202) |
2011 |
Jan
(321) |
Feb
(413) |
Mar
(299) |
Apr
(215) |
May
(284) |
Jun
(203) |
Jul
(207) |
Aug
(314) |
Sep
(321) |
Oct
(259) |
Nov
(347) |
Dec
(209) |
2012 |
Jan
(322) |
Feb
(414) |
Mar
(377) |
Apr
(179) |
May
(173) |
Jun
(234) |
Jul
(295) |
Aug
(239) |
Sep
(276) |
Oct
(355) |
Nov
(144) |
Dec
(108) |
2013 |
Jan
(170) |
Feb
(89) |
Mar
(204) |
Apr
(133) |
May
(142) |
Jun
(89) |
Jul
(160) |
Aug
(180) |
Sep
(69) |
Oct
(136) |
Nov
(83) |
Dec
(32) |
2014 |
Jan
(71) |
Feb
(90) |
Mar
(161) |
Apr
(117) |
May
(78) |
Jun
(94) |
Jul
(60) |
Aug
(83) |
Sep
(102) |
Oct
(132) |
Nov
(154) |
Dec
(96) |
2015 |
Jan
(45) |
Feb
(138) |
Mar
(176) |
Apr
(132) |
May
(119) |
Jun
(124) |
Jul
(77) |
Aug
(31) |
Sep
(34) |
Oct
(22) |
Nov
(23) |
Dec
(9) |
2016 |
Jan
(26) |
Feb
(17) |
Mar
(10) |
Apr
(8) |
May
(4) |
Jun
(8) |
Jul
(6) |
Aug
(5) |
Sep
(9) |
Oct
(4) |
Nov
|
Dec
|
2017 |
Jan
(5) |
Feb
(7) |
Mar
(1) |
Apr
(5) |
May
|
Jun
(3) |
Jul
(6) |
Aug
(1) |
Sep
|
Oct
(2) |
Nov
(1) |
Dec
|
2018 |
Jan
|
Feb
|
Mar
|
Apr
(1) |
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
2020 |
Jan
|
Feb
|
Mar
|
Apr
|
May
(1) |
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
2025 |
Jan
(1) |
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
S | M | T | W | T | F | S |
---|---|---|---|---|---|---|
|
|
|
|
1
(5) |
2
(24) |
3
(2) |
4
|
5
(4) |
6
(1) |
7
(22) |
8
(9) |
9
(23) |
10
(6) |
11
|
12
(7) |
13
(17) |
14
(12) |
15
(10) |
16
(22) |
17
(11) |
18
(6) |
19
(11) |
20
(13) |
21
(10) |
22
(7) |
23
(7) |
24
(5) |
25
(3) |
26
(20) |
27
(17) |
28
(25) |
29
(16) |
30
(17) |
31
(8) |
On 1/26/07, Fernando Perez <fpe...@gm...> wrote: > On 1/25/07, Alan G Isaac <ai...@am...> wrote: > > On 2007年1月24日, Fernando Perez apparently wrote: > > > Let us know if this is not enough or if you have any other issues. > > > > How about for Windows users? You list as dependencies: [...] > We obviously need to update the windows documentation... Done in SVN, thanks for reporting this. regards, f
On Mar 18, 2007, at 12:41 PM, John Hunter wrote: > On 3/17/07, Simson Garfinkel <si...@ac...> wrote: >> Hi. I haven't been active for a while, but now I have another paper >> that I need to get out... > > Glad to have you back... Thanks. I've taken a new job, moved to california, and have been flying between the two coasts every week. It doesn't leave much time for mailing lists... > >> Anyway, I need to draw a cumulative distribution function, as the >> reviewers of my last paper really nailed me to the wall for including >> histograms instead of CDFs. Is there any way to plot a CDF with >> matplotlib? > > For analytic cdfs, see scipy.stats. I assume you need an empirical > cdf. You can use matplotlib.mlab.hist to compute the empirical pdf > (use normed=True to return a PDF rather than a frequency count). Then > use numpy.cumsum to do the cumulative sum of the pdf, multiplying by > the binsize so it approximates the integral. > > import matplotlib.mlab > from pylab import figure, show, nx > > x = nx.mlab.randn(10000) > p,bins = matplotlib.mlab.hist(x, 50, normed=True) > db = bins[1]-bins[0] > cdf = nx.cumsum(p*db) > > fig = figure() > ax = fig.add_subplot(111) > ax.bar(bins, cdf, width=0.8*db) > show() > Thanks! I'll try it out and see what happens.
On 3/18/07, Allan Noriel Estrella <all...@gm...> wrote: > I have an array of data that were sampled with a sampling rate of 1.5625 > samples/sec . I want to plot these data with the x axis showing time in > terms of hours for the major ticks and minutes for the minor ticks in my > custom made wx App. I have been fiddling around with the HourLocator and > MinuteLocator classes but it seems I can't get them to work. Do you know an > easy way to do this? I just want to plot my data in terms of the actual > time/duration of logging they represent (in hours and/or in minutes when the > plot is zoomed in or there is less than an hour's worth of samples) Well, the HourLocator, MiinuteLocator, etc are for times and not durations. If you want durations, just use plot and not plot_date. Does that help?
On 3/17/07, Simson Garfinkel <si...@ac...> wrote: > Hi. I haven't been active for a while, but now I have another paper > that I need to get out... Glad to have you back... > Anyway, I need to draw a cumulative distribution function, as the > reviewers of my last paper really nailed me to the wall for including > histograms instead of CDFs. Is there any way to plot a CDF with > matplotlib? For analytic cdfs, see scipy.stats. I assume you need an empirical cdf. You can use matplotlib.mlab.hist to compute the empirical pdf (use normed=True to return a PDF rather than a frequency count). Then use numpy.cumsum to do the cumulative sum of the pdf, multiplying by the binsize so it approximates the integral. import matplotlib.mlab from pylab import figure, show, nx x = nx.mlab.randn(10000) p,bins = matplotlib.mlab.hist(x, 50, normed=True) db = bins[1]-bins[0] cdf = nx.cumsum(p*db) fig = figure() ax = fig.add_subplot(111) ax.bar(bins, cdf, width=0.8*db) show()
I have an array of data that were sampled with a sampling rate of 1.5625samples/sec . I want to plot these data with the x axis showing time in terms of hours for the major ticks and minutes for the minor ticks in my custom made wx App. I have been fiddling around with the HourLocator and MinuteLocator classes but it seems I can't get them to work. Do you know an easy way to do this? I just want to plot my data in terms of the actual time/duration of logging they represent (in hours and/or in minutes when the plot is zoomed in or there is less than an hour's worth of samples)
1.please compare figure 1 and figure 2 2 How to change text color on Pie chart ? (If fragment of Pie chart is black text is unreadable ) 3. Text "0.5%" in figure 2 is unreadable, how correct this? ( i try use pctdistance= 1.1 but if value is format autopct='%1.4f%%' some text is unreadable ) #!/usr/bin/env python from pylab import * # make a square figure and axes figure(1, figsize=(8,8)) ax = axes([0.1, 0.1, 0.8, 0.8]) labels = 'Frogs', 'Hogs', 'Dogs', 'Logs' fracs = [50,25,24.5, 0.5] figure(1) pie(fracs, labels=labels) legend( loc='best', shadow=True) # figure(2) show a some optional features. autopct is used to label # the percentage of the pie, and can be a format string or a function # which takes a percentage and returns a string. explode is a # len(fracs) sequence which gives the fraction of the radius to # offset that slice. figure(2, figsize=(8,8)) explode=(0, 0.05, 0, 0) pie(fracs, explode=explode, labels=labels, autopct='%1.1f%%', shadow=True) legend( loc='best', shadow=True) savefig('pie_demo') show()