Search Results

Search found 15 results on 1 pages for 'mlissner'.

Page 1/1 | 1 

  • How to debug suspend?

    - by mlissner
    I've been using Ubuntu for about five years now, and I still can't make it suspend when I want to. It's quite irritating that I can program up a storm, hack the machine in numerous other ways, and yet, and yet when I try to make it suspend or debug suspend, I fail miserably. I need help. Where do I begin to find the problem? What do I do to fix it? I'm placing a bounty on this, because I've literally lost hours of my life to this problem, and leaving my computer on ALL the time is terrible. The symptoms: Pressing suspend brings my computer to a state where it has a blinking cursor, the fans are running, it seems that the HD has turned off (I think), and I can't do anything to bring it back from this state (short of a hard reboot). Possibly related: My fans stay on even after a shutdown, and even then, I have to press the power button for five seconds before I can start it up again. I don't know what logs to look at to debug the problem, and I imagine they'd get nuked on reboot anyway. Please, please help. This drives me completely nuts, and I've been living with it for over a year.

    Read the article

  • Oneiric is freezing. Need help diagnosing and filing a bug

    - by mlissner
    Six months ago, I bought a new Sandy Bridge CPU and built myself a nice desktop machine. Until now it hasn't worked...at all. I finally have gotten it working now that Oneiric is released, but it freezes every so often, making it little more than a semi-functional temptation. What happens when the system freezes is: the music I have playing enters into about 5s loops. SSH fails the monitor freezes the mouse freezes the keyboard fails The only way to fix it is to do a hard reset...and that sucks. I'd love to at least figure out the source of the freeze so I can file a bug. I've looked in dmesg, kern.log, and the X.org logs. Nothing interesting is any of them. Since SSH fails, I'm convinced it's not an X crash: https://wiki.ubuntu.com/X/Troubleshooting/Freeze Anything else I can do?

    Read the article

  • How do I delete the Windows Explorer address bar history

    - by mlissner
    Note: I am NOT referring to Internet Explorer. I am using Windows XP and Windows Server 2008 and need to delete the history values from the file browser (aka Windows Explorer). Somebody put a password into the address bar as ftp://user:pass, and now I can't delete the value. Some forums say to delete this registry key: HKEY_CURRENT_USER\Software\Microsoft\Windows\Curre ntVersion\Explorer\TypedPaths In Windows XP, the key doesn't exist, and in Windows Server 2008, the key is there, but deleting it doesn't seem to help. Any help?

    Read the article

  • Transferring Postfix install to new computer

    - by mlissner
    I have postfix installed on one computer, with DKIM and SPF working properly. What I'd like to do is start using a different computer instead, with the minimal amount of fuss. Mail servers have a way of baffling me, but I know there are things with cryptography going on here that I don't fully understand (and I don't really care to - I figured it out when I set up the last computer about a year ago, and am happy not to delve into it again). Right now, I'm working on the early steps of this process -- installing postfix on the new machine, and getting it going. Are there specific steps I could take to move the correct configs and key files and such to the new computer?

    Read the article

  • Generating a list of values a regex COULD match in Python

    - by mlissner
    I'm trying to use a regex as an input, and from there generate all the possible values that the regex would match. So, for example, if the regex is "three-letter words starting with a, and ending in c," then the code would generate a list with the values [aac, abc, acc, adc, a1c....]. Is there an easy way to do this? I'm using python.

    Read the article

  • How do you set the sitemap priority for flatpages in django?

    - by mlissner
    I have a site with about 60,000 pages that are getting placed in the sitemap, and which have a priority of 0.3. These are all really long pages that are rich in keywords. I also have a few pages (like the about page), that need high priority, but which I've implemented with the django flatpages framework. Is it possible for pages created this way to have a higher priority in the sitemap?

    Read the article

  • How do you access/configure summaries/snippets in Django Haystack

    - by mlissner
    I'm working on getting django-haystack set up on my site, and am trying to have snippets in my search results roughly like so: Title of result one about Wikis ...this special thing about wiki values is that...I always use a wiki when I walk...snippet value three talks about wikis too...and here's another snippet value about wikis. I know there's a template tag for highlighting, but how do you generate the snippets themselves? I know Solr generates these, but I can't figure out how to get them from Haystack.

    Read the article

  • How do you set the initial value for a ManyToMany field in django?

    - by mlissner
    I am using a ModelForm to create a form, and I have gotten the initial values set for every field in the form except for the one that is a ManyToMany field. I understand that I need to give it a list, but I can't get it to work. My code right now is roughly: contacts = userProfile.contact.all() initial = {'contacts': contacts} But that doesn't work. Am I missing something here?

    Read the article

  • How to efficiently serve massive sitemaps in django

    - by mlissner
    I have a site with about 150K pages in its sitemap. I'm using the sitemap index generator to make the sitemaps, but really, I need a way of caching it, because building the 150 sitemaps of 1,000 links each is brutal on my server.[1] I COULD cache each of these sitemap pages with memcached, which is what I'm using elsewhere on the site...however, this is so many sitemaps that it would completely fill memcached....so that doesn't work. What I think I need is a way to use the database as the cache for these, and to only generate them when there are changes to them (which as a result of the sitemap index means only changing the latest couple of sitemap pages, since the rest are always the same.)[2] But, as near as I can tell, I can only use one cache backend with django. How can I have these sitemaps ready for when Google comes-a-crawlin' without killing my database or memcached? Any thoughts? [1] I've limited it to 1,000 links per sitemap page because generating the max, 50,000 links, just wasn't happening. [2] for example, if I have sitemap.xml?page=1, page=2...sitemap.xml?page=50, I only really need to change sitemap.xml?page=50 until it is full with 1,000 links, then I can it pretty much forever, and focus on page 51 until it's full, cache it forever, etc.

    Read the article

  • After extending the User model in django, how do you create a ModelForm?

    - by mlissner
    I extended the User model in django to include several other variables, such as location, and employer. Now I'm trying to create a form that has the following fields: First name (from User) Last name (from User) Location (from UserProfile, which extends User via a foreign key) Employer (also from UserProfile) I have created a modelform: from django.forms import ModelForm from django.contrib import auth from alert.userHandling.models import UserProfile class ProfileForm(ModelForm): class Meta: # model = auth.models.User # this gives me the User fields model = UserProfile # this gives me the UserProfile fields So, my question is, how can I create a ModelForm that has access to all of the fields, whether they are from the User model or the UserProfile model? Hope this makes sense. I'll be happy to clarify if there are any questions.

    Read the article

  • MySQL efficiency as it relates to the database/table size

    - by mlissner
    I'm building a system using django, Sphinx and MySQL that's very quickly becoming quite large. The database currently has about 2000 rows, and I've written a program that's going to populate it with another 40,000 rows in a couple days. Since the database is live right now, and since I've never had a database with this much information in it, I'm worried about some things: Is adding all these rows going to seriously degrade the efficiency of my django app? Will I need to go back through it and optimize all my database calls so they're doing things more cleverly? Or will this make the database slow all around to the extent that I can't do anything about it at all? If you scoff at my 40k rows, then, my next question is, at what point SHOULD I be concerned? I will likely be adding another couple hundred thousand soon, so I worry, and I fret. How is sphinx going to feel about all this? Is it going to freak out when it realizes it has to index all this data? Or will it be fine? Is this normal for it? If it is, at what point should I be concerned that it's too much data for Sphinx? Thanks for any thoughts.

    Read the article

  • Regex for finding valid sphinx fields

    - by mlissner
    I'm trying to validate that the fields given to sphinx are valid, but I'm having difficulty. Imagine that valid fields are cat, mouse, dog, puppy. Valid searches would then be: @cat search terms @(cat) search terms @(cat, dog) search term @cat searchterm1 @dog searchterm2 @(cat, dog) searchterm1 @mouse searchterm2 So, I want to use a regular expression to find terms such as cat, dog, mouse in the above examples, and check them against a list of valid terms. Thus, a query such as: @(goat) Would produce an error because goat is not a valid term. I've gotten so that I can find simple queries such as @cat with this regex: (?:@)([^( ]*) But I can't figure out how to find the rest. I'm using python & django, for what that's worth.

    Read the article

  • How to parse multiple dates from a block of text in Python (or another language)

    - by mlissner
    I have a string that has several date values in it, and I want to parse them all out. The string is natural language, so the best thing I've found so far is dateutil. Unfortunately, if a string has multiple date values in it, dateutil throws an error: >>> s = "I like peas on 2011-04-23, and I also like them on easter and my birthday, the 29th of July, 1928" >>> parse(s, fuzzy=True) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/lib/pymodules/python2.7/dateutil/parser.py", line 697, in parse return DEFAULTPARSER.parse(timestr, **kwargs) File "/usr/lib/pymodules/python2.7/dateutil/parser.py", line 303, in parse raise ValueError, "unknown string format" ValueError: unknown string format Any thoughts on how to parse all dates from a long string? Ideally, a list would be created, but I can handle that myself if I need to. I'm using Python, but at this point, other languages are probably OK, if they get the job done. PS - I guess I could recursively split the input file in the middle and try, try again until it works, but it's a hell of a hack.

    Read the article

1