* contrib/babel/library-of-babel.org(read): adding explicit format
argument
(gdoc-read): explicit format argument, and passing the csv
specification through to org-table-import
also removing org example code blocks
* lisp/ob-lob.el (org-babel-lob-execute): changing indentation to
improve line length
(org-babel-lob-get-info): now catching indexes passed through lob
calls
(org-babel-lob-one-liner-regexp): now matches optional indexes into
variable results
OK, we finally have a make file where I can do a complete release,
including all the git updates on master and maint branches, with a
single command:
make release TAG=N.NN
It is important that the working directory is clean when calling this
command.
Since this command can fail and then leave the git repo in an
uncertain state, it is best to first run this:
make testrelease
This will create a throw-away branch "testrelease" and try alll the
commands there, to make sure everything is fine.
The workflow is now:
$ pw branch NNN
To create a branch with patch numb er NNN
$ git commit --amend
To check and modify the commit message where necessary
$ pw merge -m "Comment for notification email" NNN
To merge that branch. The -m comment will be added to the
notification email that is sent to the mailing list.
The -m comment can also be given to the update command,if the update
sends an email:
$ pw update -s "changes requested" -m "comment for email" NNN
* lisp/org-table.el (org-table-convert-region): don't continue csv
importation which the point catches the end, this fixes an infinite
loop which was caused by the (point) never catching up with the
"end" marker
Hi,
I've just noticed that I get an infinite loop when importing csv tables
into Org-mode tables. For example calling
(org-table-convert-region (point-min) (point-max) '(4))
from inside of a simple text file containing the following
--8<---------------cut here---------------start------------->8---
1,2,3
1
2
3
5
--8<---------------cut here---------------end--------------->8---
results in an infinite loop. The attached patch fixes this issue [1],
however it seems weird that this would just surface now, given that the
code in question has been in the repo since last fall
,----
| commit 59c9c4cdd4
| Author: Carsten Dominik <carsten.dominik@gmail.com>
| Date: Mon Oct 26 12:31:16 2009 +0100
|
| Correctly interpret CVS tables with quoted fields
|
| The csv parser was very primitive, ignoring quoted fields. This is
| now fixed.
`----
Has anyone else noticed this problem? Should I go ahead and apply this
patch?
Best -- Eric
Footnotes:
[1]
>From 3c3f4ca9a34dca23051ca2f4e4518b416338d4f4 Mon Sep 17 00:00:00 2001
From: Eric Schulte <schulte.eric@gmail.com>
Date: Mon, 19 Jul 2010 13:45:25 -0700
Subject: [PATCH] org-table: fix infinite loop in table importation
* lisp/org-table.el (org-table-convert-region): don't continue csv
importation which the point catches the end, this fixes an infinite
loop which was caused by the (point) never catching up with the
"end" marker
* lisp/org-macs.el (org-string-nw-p): New function.
* lisp/org-capture.el (org-capture-import-remember-templates): Interpret
an empty string as request to use `org-default-notes-file'.
(org-capture-target-buffer): If the FILE is not a (non-empty) string,
use `org-default-notes-file'.
* lisp/org-colview-xemacs.el (org-columns-compile-map):
(org-columns-number-to-string):
(org-columns-string-to-number): Handle estimate ranges.
(org-estimate-mean-and-var): New function.
(org-estimate-combine): New function.
(org-estimate-print): New function.
(org-string-to-estimate): New function.
Carsten,
Here is a patch for a new 'est+' summary type, including corresponding changes for xemacs and the manual. I've done basic testing on the GNU emacs version, but not the xemacs code.
I'm not sure the change to the manual provides the right amount of information
in the right place; I'd be happy to re-write to make it find in
better. Similarly, the name of the summary type is entirely up to you.
I didn't know whether to send this directly to you, or to the list; if it should go to the list I'd be happy to send it there directly.
Thanks for the chance to contribute,
Mike