< Return to Video

Dalle-2 Is Now Open to Everyone! - Getting Started Tutorial

  • 0:00 - 0:01
    Hey, everybody.
  • 0:01 - 0:02
    Scott Detweiler here.
  • 0:02 - 0:03
    So, the big news today is that DALL-E 2 is
  • 0:03 - 0:04
    now open and available
  • 0:04 - 0:06
    for everybody to try it.
  • 0:06 - 0:07
    So, if you've been on that waiting list
  • 0:07 - 0:08
    and you haven't been called,
  • 0:08 - 0:10
    now's your chance to play with it anyway.
  • 0:10 - 0:12
    So, unlike Stable Diffusion or Midjourney,
  • 0:12 - 0:13
    some of the other AI art tools out there,
  • 0:13 - 0:15
    it has been around a while.
  • 0:15 - 0:17
    It existed earlier as Craiyon or
  • 0:17 - 0:20
    DALL-E Mini, which isn't as powerful.
  • 0:20 - 0:21
    DALL-E is a little bit
  • 0:21 - 0:23
    more stock photography looking
  • 0:23 - 0:25
    if you're looking for a description
  • 0:25 - 0:26
    of what it produces, say Midjourney,
  • 0:26 - 0:27
    which is more painterly,
  • 0:27 - 0:28
    Stable Diffusion, which seems
  • 0:28 - 0:30
    to do pretty much everything.
  • 0:30 - 0:31
    For example, if you need a picture
  • 0:31 - 0:32
    of a cat with a hand grenade,
  • 0:32 - 0:33
    you can get one.
  • 0:33 - 0:35
    It's exactly that type of tool,
  • 0:35 - 0:37
    but it has some other really cool
  • 0:37 - 0:39
    capabilities, and that's the ability to go
  • 0:39 - 0:41
    in and modify a part of an image and give
  • 0:41 - 0:43
    it a very basic prompt, and it will
  • 0:43 - 0:45
    be able to fix that area of the image.
  • 0:45 - 0:46
    So, I use it more for troubleshooting
  • 0:46 - 0:48
    in my photography than I
  • 0:48 - 0:49
    do for AI art generation.
  • 0:49 - 0:51
    So, let's take a quick look at it so
  • 0:51 - 0:52
    you can kind of see how to use it.
  • 0:52 - 0:53
    I'm gonna give you some basics,
  • 0:53 - 0:55
    and then you can go and play.
  • 0:55 - 0:56
    Now the DALL-E 2 interface
  • 0:56 - 0:57
    is pretty simple.
  • 0:57 - 0:58
    It's just web-based,
  • 0:58 - 1:00
    and you've got just your prompt.
  • 1:00 - 1:01
    No other settings really,
  • 1:01 - 1:02
    so you're basically just
  • 1:02 - 1:04
    gonna type in what you'd like.
  • 1:04 - 1:05
    One of the refreshing things about
  • 1:05 - 1:06
    this system is that it
  • 1:06 - 1:07
    is very good at English.
  • 1:08 - 1:10
    So, it understands sentence structure,
  • 1:10 - 1:12
    unlike Midjourney or Stable Diffusion
  • 1:12 - 1:14
    where you've kind of got to comma separate
  • 1:14 - 1:15
    things or space separate things,
  • 1:15 - 1:16
    and it really doesn't matter
  • 1:16 - 1:17
    how you put it in there.
  • 1:17 - 1:18
    This one actually
  • 1:18 - 1:20
    is very good at sentence structure.
  • 1:20 - 1:21
    Another thing is that this seems to be
  • 1:21 - 1:23
    trained more with stock photography,
  • 1:23 - 1:24
    so you're going to get more
  • 1:24 - 1:26
    of a photographic look from this.
  • 1:26 - 1:27
    At least that's how
  • 1:27 - 1:27
    I've been looking at it.
  • 1:27 - 1:29
    When I'm looking to make art,
  • 1:29 - 1:30
    I'm reaching for Midjourney
  • 1:30 - 1:30
    or Stable Diffusion.
  • 1:31 - 1:32
    This is more of a problem-solving tool
  • 1:32 - 1:34
    for me, and a very good one at that,
  • 1:34 - 1:36
    but it is not one that I use
  • 1:36 - 1:37
    for generating art from scratch.
  • 1:37 - 1:38
    So, for example, you can ask
  • 1:38 - 1:40
    it pretty much for anything.
  • 1:40 - 1:41
    A bag of potato chips that is on fire.
  • 1:41 - 1:44
    Notice that every generation inside
  • 1:44 - 1:46
    of here will cost you a credit.
  • 1:46 - 1:47
    Up at the top, you have
  • 1:47 - 1:48
    access to your membership.
  • 1:48 - 1:50
    You can see how many credits you have.
  • 1:50 - 1:51
    So, this did not cost us four credits, but
  • 1:51 - 1:53
    it cost us one for each generation shot.
  • 1:53 - 1:54
    But the most exciting part
  • 1:55 - 1:56
    of this tool is not this,
  • 1:56 - 1:57
    it's actually the outpainting.
  • 1:57 - 1:59
    So, if you click up here and you go
  • 1:59 - 2:01
    to Outpainting, it's gonna open an editor,
  • 2:01 - 2:03
    this is the part that is amazing to me.
  • 2:04 - 2:05
    We've got two options.
  • 2:05 - 2:07
    You can either work with an image
  • 2:07 - 2:08
    and generate pieces of it,
  • 2:08 - 2:10
    or you can start and put a piece
  • 2:10 - 2:12
    in the middle and generate outwards.
  • 2:12 - 2:13
    I'll show you each one.
  • 2:13 - 2:15
    So, the generation frame in here is 1,024
  • 2:15 - 2:17
    pixels by 1,024 pixels,
  • 2:17 - 2:19
    and we can load an image on top of it.
  • 2:19 - 2:20
    So, I'm gonna load one in here,
  • 2:20 - 2:21
    and notice that this image
  • 2:21 - 2:22
    is larger than 1,000 pixels.
  • 2:23 - 2:24
    So, if I want to, I can have
  • 2:24 - 2:25
    the generation frame come in.
  • 2:25 - 2:27
    We can work our way around the image, or
  • 2:28 - 2:29
    I can scale the image down
  • 2:30 - 2:32
    so that it's under the 1,024.
  • 2:32 - 2:33
    In this case, we'll make it
  • 2:33 - 2:34
    a little bit farther under.
  • 2:34 - 2:37
    When I click OK, I'll be handed this back.
  • 2:37 - 2:38
    Now hold down your space bar,
  • 2:38 - 2:40
    you can actually move around the Canvas.
  • 2:40 - 2:42
    The interface is a little wonky.
  • 2:42 - 2:43
    Like I really don't like this white bar
  • 2:43 - 2:45
    that comes across, because if we zoom in,
  • 2:45 - 2:46
    for example, at some point the white
  • 2:46 - 2:48
    bar cuts through the image.
  • 2:48 - 2:50
    You can also use your control key
  • 2:50 - 2:50
    and your mouse wheel if you'd
  • 2:50 - 2:51
    like to scroll this way.
  • 2:53 - 2:54
    And then you can move the generation frame
  • 2:54 - 2:56
    by left-clicking inside of it anywhere.
  • 2:56 - 2:57
    So, we can position this
  • 2:57 - 2:59
    image where we would like.
  • 3:00 - 3:01
    Now again, because it understands English
  • 3:01 - 3:02
    pretty well, we don't have
  • 3:02 - 3:04
    to overemphasize the prompt here.
  • 3:04 - 3:05
    Let's type in autumn
  • 3:05 - 3:07
    forest and hit generate.
  • 3:08 - 3:10
    This will cost us one generation,
  • 3:10 - 3:11
    and we'll be given four options.
  • 3:14 - 3:15
    As you can see, it does a very good job
  • 3:15 - 3:17
    of matching the style and color
  • 3:17 - 3:18
    tones of the image for the most part.
  • 3:18 - 3:21
    Sometimes there's a few that are special.
  • 3:24 - 3:25
    These are pretty good.
  • 3:26 - 3:27
    Once you find one you like,
  • 3:27 - 3:28
    just download it.
  • 3:30 - 3:31
    Now if we refresh this by clicking
  • 3:31 - 3:32
    on the name up here,
  • 3:32 - 3:34
    we can go back up to show editor.
  • 3:36 - 3:37
    Now we could try it the other way.
  • 3:37 - 3:38
    For example, let's work
  • 3:38 - 3:39
    with this image here.
  • 3:40 - 3:41
    So, you can see it's much larger
  • 3:41 - 3:43
    than 1,024 pixels, but that's okay,
  • 3:43 - 3:45
    we can accept it, and now we have our
  • 3:45 - 3:46
    generation frame, and we can float it.
  • 3:46 - 3:48
    So, we can move it off to one side,
  • 3:48 - 3:49
    for example, if we wish
  • 3:49 - 3:50
    to continue this image.
  • 4:31 - 4:35
    You can see it, did a pretty nice job of trying to match the context what was given
  • 4:36 - 4:37
    with an extremely small prompt, which I think is really impressive
  • 4:37 - 4:40
    Lastly, what you can do is you can choose to use the Eraser
  • 4:41 - 4:44
    When you use the Eraser, you can take part of the image out, example here,
  • 4:44 - 4:47
    we remove this window.
  • 4:51 - 4:53
    Uber generation frame there, click generate again.
  • 5:02 - 5:07
    But the coolest things you can do with the outpinning tool is put an image in, have it generate the outside and then load the image in again
  • 5:07 - 5:16
    and again creating what looks like an infinite zoom out, are you zooming out of the image because each time you're asking it to generate the perimeter.
  • 5:15 - 5:26
    We can start with this image for example. I just need to make sure that I'm less than 1,024 pixels on the long edge. I can position the generation frame where
  • 5:26 - 5:28
    I'd like it and then continue working.
  • 5:38 - 5:39
    Again, oftentimes, an image will be
  • 5:39 - 5:41
    a little bit off, but that's okay.
  • 5:41 - 5:42
    We get four options,
  • 5:42 - 5:44
    and we only pay for one.
  • 5:46 - 5:47
    Now what I would do is I would download
  • 5:47 - 5:48
    this, and I would re-upload
  • 5:48 - 5:50
    it and start from here.
  • 5:51 - 5:52
    So, that's DALL-E 2,
  • 5:52 - 5:53
    very powerful.
  • 5:53 - 5:54
    As I said, I use it more
  • 5:54 - 5:55
    for troubleshooting than I did
  • 5:55 - 5:58
    for ideation or art generation.
  • 5:58 - 6:01
    But it's pretty much the most amazing
  • 6:01 - 6:02
    stock photography generator that it's ever
  • 6:02 - 6:05
    going to be and a great way to
  • 6:06 - 6:07
    edit parts of an image that seem like
  • 6:07 - 6:08
    they're almost impossible to find
  • 6:08 - 6:10
    something that's gonna fit.
  • 6:10 - 6:11
    It fits perfectly.
  • 6:11 - 6:12
    So, I'm very excited.
  • 6:12 - 6:13
    I'm glad that it's public now.
  • 6:13 - 6:15
    I think it is, because Stable Diffusion
  • 6:15 - 6:16
    and Midjourney have both been very
  • 6:16 - 6:18
    popular, and it's been very difficult
  • 6:18 - 6:19
    to get access to the system.
  • 6:19 - 6:20
    So, I'm glad to see that they
  • 6:20 - 6:21
    have finally opened that up.
  • 6:22 - 6:23
    So, give it a whirl and let me know
  • 6:23 - 6:24
    what you think in the comments below.
  • 6:24 - 6:25
    Everybody take care.
  • 6:25 - 6:27
    Stay safe, and I'll catch you all next time.
Title:
Dalle-2 Is Now Open to Everyone! - Getting Started Tutorial
Description:

more » « less
Video Language:
English
Team:
BYU Continuing Education
Project:
IP &amp; T-371(BYUO)
Duration:
06:32

English subtitles

Revisions